The NYA’s Offline Speech Translation System for IWSLT 2024

Yingxin Zhang, Guodong Ma, Binbin Du


Abstract
This paper reports the NYA’s submissions to IWSLT 2024 Offline Speech Translation (ST) task on the sub-tasks including English to Chinese, Japanese, and German. In detail, we participate in the unconstrained training track using the cascaded ST structure. For the automatic speech recognition (ASR) model, we use the Whisper large-v3 model. For the neural machine translation (NMT) model, the wider and deeper Transformer is adapted as the backbone model. Furthermore, we use data augmentation technologies to augment training data and data filtering strategies to improve the quality of training data. In addition, we explore many MT technologies such as Back Translation, Forward Translation, R-Drop, and Domain Adaptation.
Anthology ID:
2024.iwslt-1.6
Volume:
Proceedings of the 21st International Conference on Spoken Language Translation (IWSLT 2024)
Month:
August
Year:
2024
Address:
Bangkok, Thailand (in-person and online)
Editors:
Elizabeth Salesky, Marcello Federico, Marine Carpuat
Venue:
IWSLT
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
39–45
Language:
URL:
https://aclanthology.org/2024.iwslt-1.6
DOI:
Bibkey:
Cite (ACL):
Yingxin Zhang, Guodong Ma, and Binbin Du. 2024. The NYA’s Offline Speech Translation System for IWSLT 2024. In Proceedings of the 21st International Conference on Spoken Language Translation (IWSLT 2024), pages 39–45, Bangkok, Thailand (in-person and online). Association for Computational Linguistics.
Cite (Informal):
The NYA’s Offline Speech Translation System for IWSLT 2024 (Zhang et al., IWSLT 2024)
Copy Citation:
PDF:
https://aclanthology.org/2024.iwslt-1.6.pdf