A Study of Multilingual versus Meta-Learning for Language Model Pre-Training for Adaptation to Unseen Low Resource Languages

Jyotsana Khatri, Rudra Murthy, Amar Prakash Azad, Pushpak Bhattacharyya


Abstract
In this paper, we compare two approaches to train a multilingual language model: (i) simple multilingual learning using data-mixing, and (ii) meta-learning. We examine the performance of these models by extending them to unseen language pairs and further finetune them for the task of unsupervised NMT. We perform several experiments with varying amounts of data and give a comparative analysis of the approaches. We observe that both approaches give a comparable performance, and meta-learning gives slightly better results in a few cases of low amounts of data. For Oriya-Punjabi language pair, meta-learning performs better than multilingual learning when using 2M, and 3M sentences.
Anthology ID:
2023.mtsummit-research.3
Volume:
Proceedings of Machine Translation Summit XIX, Vol. 1: Research Track
Month:
September
Year:
2023
Address:
Macau SAR, China
Editors:
Masao Utiyama, Rui Wang
Venue:
MTSummit
SIG:
Publisher:
Asia-Pacific Association for Machine Translation
Note:
Pages:
26–34
Language:
URL:
https://aclanthology.org/2023.mtsummit-research.3
DOI:
Bibkey:
Cite (ACL):
Jyotsana Khatri, Rudra Murthy, Amar Prakash Azad, and Pushpak Bhattacharyya. 2023. A Study of Multilingual versus Meta-Learning for Language Model Pre-Training for Adaptation to Unseen Low Resource Languages. In Proceedings of Machine Translation Summit XIX, Vol. 1: Research Track, pages 26–34, Macau SAR, China. Asia-Pacific Association for Machine Translation.
Cite (Informal):
A Study of Multilingual versus Meta-Learning for Language Model Pre-Training for Adaptation to Unseen Low Resource Languages (Khatri et al., MTSummit 2023)
Copy Citation:
PDF:
https://aclanthology.org/2023.mtsummit-research.3.pdf