Alberto Cetoli
2022
Dispatcher: A Message-Passing Approach to Language Modelling
Alberto Cetoli
Proceedings of the 2022 CLASP Conference on (Dis)embodiment
This paper proposes a message-passing mechanism to address language modelling. A new layer type is introduced that aims to substitute self-attention for unidirectional sequence generation tasks. The system is shown to be competitive with existing methods: Given N tokens, the computational complexity is O(N logN) and the memory complexity is O(N) under reasonable assumptions. In the end, the Dispatcher layer is seen to achieve comparable perplexity to self-attention while being more efficient.
2020
Exploring the zero-shot limit of FewRel
Alberto Cetoli
Proceedings of the 28th International Conference on Computational Linguistics
This paper proposes a general purpose relation extractor that uses Wikidata descriptions to represent the relation’s surface form. The results are tested on the FewRel 1.0 dataset, which provides an excellent framework for training and evaluating the proposed zero-shot learning system in English. This relation extractor architecture exploits the implicit knowledge of a language model through a question-answering approach.
2017
Graph Convolutional Networks for Named Entity Recognition
Alberto Cetoli
|
Stefano Bragaglia
|
Andrew O’Harney
|
Marc Sloan
Proceedings of the 16th International Workshop on Treebanks and Linguistic Theories
Search