Sub-Word Alignment is Still Useful: A Vest-Pocket Method for Enhancing Low-Resource Machine Translation

Minhan Xu, Yu Hong


Abstract
We leverage embedding duplication between aligned sub-words to extend the Parent-Child transfer learning method, so as to improve low-resource machine translation. We conduct experiments on benchmark datasets of My-En, Id-En and Tr-En translation scenarios. The test results show that our method produces substantial improvements, achieving the BLEU scores of 22.5, 28.0 and 18.1 respectively. In addition, the method is computationally efficient which reduces the consumption of training time by 63.8%, reaching the duration of 1.6 hours when training on a Tesla 16GB P100 GPU. All the models and source codes in the experiments will be made publicly available to support reproducible research.
Anthology ID:
2022.acl-short.68
Volume:
Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers)
Month:
May
Year:
2022
Address:
Dublin, Ireland
Editors:
Smaranda Muresan, Preslav Nakov, Aline Villavicencio
Venue:
ACL
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
613–619
Language:
URL:
https://aclanthology.org/2022.acl-short.68
DOI:
10.18653/v1/2022.acl-short.68
Bibkey:
Cite (ACL):
Minhan Xu and Yu Hong. 2022. Sub-Word Alignment is Still Useful: A Vest-Pocket Method for Enhancing Low-Resource Machine Translation. In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers), pages 613–619, Dublin, Ireland. Association for Computational Linguistics.
Cite (Informal):
Sub-Word Alignment is Still Useful: A Vest-Pocket Method for Enhancing Low-Resource Machine Translation (Xu & Hong, ACL 2022)
Copy Citation:
PDF:
https://aclanthology.org/2022.acl-short.68.pdf
Software:
 2022.acl-short.68.software.zip
Code
 Cosmos-Break/transfer-mt-submit