2022
pdf
bib
Pre-training Language Models for Surface Realization
Farhood Farahnak
|
Leila Kosseim
Proceedings of the 5th International Conference on Natural Language and Speech Processing (ICNLSP 2022)
2020
pdf
bib
abs
Surface Realization Using Pretrained Language Models
Farhood Farahnak
|
Laya Rafiee
|
Leila Kosseim
|
Thomas Fevens
Proceedings of the Third Workshop on Multilingual Surface Realisation
In the context of Natural Language Generation, surface realization is the task of generating the linear form of a text following a given grammar. Surface realization models usually consist of a cascade of complex sub-modules, either rule-based or neural network-based, each responsible for a specific sub-task. In this work, we show that a single encoder-decoder language model can be used in an end-to-end fashion for all sub-tasks of surface realization. The model is designed based on the BART language model that receives a linear representation of unordered and non-inflected tokens in a sentence along with their corresponding Universal Dependency information and produces the linear sequence of inflected tokens along with the missing words. The model was evaluated on the shallow and deep tracks of the 2020 Surface Realization Shared Task (SR’20) using both human and automatic evaluation. The results indicate that despite its simplicity, our model achieves competitive results among all participants in the shared task.
2019
pdf
bib
abs
The Concordia NLG Surface Realizer at SRST 2019
Farhood Farahnak
|
Laya Rafiee
|
Leila Kosseim
|
Thomas Fevens
Proceedings of the 2nd Workshop on Multilingual Surface Realisation (MSR 2019)
This paper presents the model we developed for the shallow track of the 2019 NLG Surface Realization Shared Task. The model reconstructs sentences whose word order and word inflections were removed. We divided the problem into two sub-problems: reordering and inflecting. For the purpose of reordering, we used a pointer network integrated with a transformer model as its encoder-decoder modules. In order to generate the inflected forms of tokens, a Feed Forward Neural Network was employed.
2018
pdf
bib
abs
CLaC @ DEFT 2018: Sentiment analysis of tweets on transport from Île-de-France
Simon Jacques
|
Farhood Farahnak
|
Leila Kosseim
Actes de la Conférence TALN. Volume 2 - Démonstrations, articles des Rencontres Jeunes Chercheurs, ateliers DeFT
CLaC @ DEFT 2018: Analysis of tweets on transport on the Île-de-France This paper describes the system deployed by the CLaC lab at Concordia University in Montreal for the DEFT 2018 shared task. The competition consisted in four different tasks; however, due to lack of time, we only participated in the first two. We participated with a system based on conventional supervised learning methods: a support vector machine classifier and an artificial neural network. For task 1, our best approach achieved an F-measure of 87.61%; while at task 2, we achieve 51.03%, situating our system below the average of the other participants.