The SETU-ADAPT Submissions to WMT 2024 Chat Translation Tasks

Maria Zafar, Antonio Castaldo, Prashanth Nayak, Rejwanul Haque, Andy Way


Abstract
This paper presents the SETU-ADAPT submissions to the WMT24 Chat Translation Task. Large language models (LLM) currently provides the state-of-the-art solutions in many natural language processing (NLP) problems including machine translation (MT). For the WMT24 Chat Translation Task we leveraged LLMs for their MT capabilities. In order to adapt the LLMs for a specific domain of interest, we explored different fine-tuning and prompting strategies. We also employed efficient data retrieval methods to curate the data used for fine-tuning. We carried out experiments for two language pairs: German-to-English and French-to-English. Our MT models were evaluated using three metrics: BLEU, chrF and COMET. In this paper we describes our experiments including training setups, results and findings.
Anthology ID:
2024.wmt-1.104
Volume:
Proceedings of the Ninth Conference on Machine Translation
Month:
November
Year:
2024
Address:
Miami, Florida, USA
Editors:
Barry Haddow, Tom Kocmi, Philipp Koehn, Christof Monz
Venue:
WMT
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
1023–1030
Language:
URL:
https://aclanthology.org/2024.wmt-1.104
DOI:
Bibkey:
Cite (ACL):
Maria Zafar, Antonio Castaldo, Prashanth Nayak, Rejwanul Haque, and Andy Way. 2024. The SETU-ADAPT Submissions to WMT 2024 Chat Translation Tasks. In Proceedings of the Ninth Conference on Machine Translation, pages 1023–1030, Miami, Florida, USA. Association for Computational Linguistics.
Cite (Informal):
The SETU-ADAPT Submissions to WMT 2024 Chat Translation Tasks (Zafar et al., WMT 2024)
Copy Citation:
PDF:
https://aclanthology.org/2024.wmt-1.104.pdf