mixSeq: A Simple Data Augmentation Methodfor Neural Machine Translation

Xueqing Wu, Yingce Xia, Jinhua Zhu, Lijun Wu, Shufang Xie, Yang Fan, Tao Qin


Abstract
Data augmentation, which refers to manipulating the inputs (e.g., adding random noise,masking specific parts) to enlarge the dataset,has been widely adopted in machine learning. Most data augmentation techniques operate on a single input, which limits the diversity of the training corpus. In this paper, we propose a simple yet effective data augmentation technique for neural machine translation, mixSeq, which operates on multiple inputs and their corresponding targets. Specifically, we randomly select two input sequences,concatenate them together as a longer input aswell as their corresponding target sequencesas an enlarged target, and train models on theaugmented dataset. Experiments on nine machine translation tasks demonstrate that such asimple method boosts the baselines by a non-trivial margin. Our method can be further combined with single input based data augmentation methods to obtain further improvements.
Anthology ID:
2021.iwslt-1.23
Volume:
Proceedings of the 18th International Conference on Spoken Language Translation (IWSLT 2021)
Month:
August
Year:
2021
Address:
Bangkok, Thailand (online)
Editors:
Marcello Federico, Alex Waibel, Marta R. Costa-jussà, Jan Niehues, Sebastian Stuker, Elizabeth Salesky
Venue:
IWSLT
SIG:
SIGSLT
Publisher:
Association for Computational Linguistics
Note:
Pages:
192–197
Language:
URL:
https://aclanthology.org/2021.iwslt-1.23
DOI:
10.18653/v1/2021.iwslt-1.23
Bibkey:
Cite (ACL):
Xueqing Wu, Yingce Xia, Jinhua Zhu, Lijun Wu, Shufang Xie, Yang Fan, and Tao Qin. 2021. mixSeq: A Simple Data Augmentation Methodfor Neural Machine Translation. In Proceedings of the 18th International Conference on Spoken Language Translation (IWSLT 2021), pages 192–197, Bangkok, Thailand (online). Association for Computational Linguistics.
Cite (Informal):
mixSeq: A Simple Data Augmentation Methodfor Neural Machine Translation (Wu et al., IWSLT 2021)
Copy Citation:
PDF:
https://aclanthology.org/2021.iwslt-1.23.pdf
Data
FLoRes