From Bilingual to Multilingual Neural Machine Translation by Incremental Training

Carlos Escolano, Marta R. Costa-jussà, José A. R. Fonollosa


Abstract
Multilingual Neural Machine Translation approaches are based on the use of task specific models and the addition of one more language can only be done by retraining the whole system. In this work, we propose a new training schedule that allows the system to scale to more languages without modification of the previous components based on joint training and language-independent encoder/decoder modules allowing for zero-shot translation. This work in progress shows close results to state-of-the-art in the WMT task.
Anthology ID:
P19-2033
Volume:
Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics: Student Research Workshop
Month:
July
Year:
2019
Address:
Florence, Italy
Editors:
Fernando Alva-Manchego, Eunsol Choi, Daniel Khashabi
Venue:
ACL
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
236–242
Language:
URL:
https://aclanthology.org/P19-2033
DOI:
10.18653/v1/P19-2033
Bibkey:
Cite (ACL):
Carlos Escolano, Marta R. Costa-jussà, and José A. R. Fonollosa. 2019. From Bilingual to Multilingual Neural Machine Translation by Incremental Training. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics: Student Research Workshop, pages 236–242, Florence, Italy. Association for Computational Linguistics.
Cite (Informal):
From Bilingual to Multilingual Neural Machine Translation by Incremental Training (Escolano et al., ACL 2019)
Copy Citation:
PDF:
https://aclanthology.org/P19-2033.pdf