GTCOM and DLUT’s Neural Machine Translation Systems for WMT23

Hao Zong


Abstract
This paper presents the submission by Global Tone Communication Co., Ltd. and Dalian Univeristy of Technology for the WMT23 shared general Machine Translation (MT) task at the Conference on Empirical Methods in Natural Language Processing (EMNLP). Our participation spans 8 language pairs, including English-Ukrainian, Ukrainian-English, Czech-Ukrainian, English-Hebrew, Hebrew-English, English-Czech, German-English, and Japanese-English. Our systems are designed without any specific constraints or requirements, allowing us to explore a wider range of possibilities in machine translation. We prioritize backtranslation, utilize multilingual translation models, and employ fine-tuning strategies to enhance performance. Additionally, we propose a novel data generation method that leverages human annotation to generate high-quality training data, resulting in improved system performance. Specifically, we use a combination of human-generated and machine-generated data to fine-tune our models, leading to more accurate translations. The automatic evaluation results show that our system ranks first in terms of BLEU score in Ukrainian-English, Hebrew-English, English-Hebrew, and German-English.
Anthology ID:
2023.wmt-1.20
Volume:
Proceedings of the Eighth Conference on Machine Translation
Month:
December
Year:
2023
Address:
Singapore
Editors:
Philipp Koehn, Barry Haddow, Tom Kocmi, Christof Monz
Venue:
WMT
SIG:
SIGMT
Publisher:
Association for Computational Linguistics
Note:
Pages:
192–197
Language:
URL:
https://aclanthology.org/2023.wmt-1.20
DOI:
10.18653/v1/2023.wmt-1.20
Bibkey:
Cite (ACL):
Hao Zong. 2023. GTCOM and DLUT’s Neural Machine Translation Systems for WMT23. In Proceedings of the Eighth Conference on Machine Translation, pages 192–197, Singapore. Association for Computational Linguistics.
Cite (Informal):
GTCOM and DLUT’s Neural Machine Translation Systems for WMT23 (Zong, WMT 2023)
Copy Citation:
PDF:
https://aclanthology.org/2023.wmt-1.20.pdf