Getting The Most Out of Your Training Data: Exploring Unsupervised Tasks for Morphological Inflection

Abhishek Purushothama, Adam Wiemerslage, Katharina Wense


Abstract
Pre-trained transformers such as BERT have been shown to be effective in many natural language tasks. However, they are under-explored for character-level sequence to sequence tasks. In this work, we investigate pre-training transformers for the character-level task of morphological inflection in several languages. We compare various training setups and secondary tasks where unsupervised data taken directly from the target task is used. We show that training on secondary unsupervised tasks increases inflection performance even without any external data, suggesting that models learn from additional unsupervised tasks themselves—not just from additional data. We also find that this does not hold true for specific combinations of secondary task and training setup, which has interesting implications for denoising objectives in character-level tasks.
Anthology ID:
2024.emnlp-main.1055
Volume:
Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing
Month:
November
Year:
2024
Address:
Miami, Florida, USA
Editors:
Yaser Al-Onaizan, Mohit Bansal, Yun-Nung Chen
Venue:
EMNLP
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
18955–18970
Language:
URL:
https://aclanthology.org/2024.emnlp-main.1055
DOI:
Bibkey:
Cite (ACL):
Abhishek Purushothama, Adam Wiemerslage, and Katharina Wense. 2024. Getting The Most Out of Your Training Data: Exploring Unsupervised Tasks for Morphological Inflection. In Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing, pages 18955–18970, Miami, Florida, USA. Association for Computational Linguistics.
Cite (Informal):
Getting The Most Out of Your Training Data: Exploring Unsupervised Tasks for Morphological Inflection (Purushothama et al., EMNLP 2024)
Copy Citation:
PDF:
https://aclanthology.org/2024.emnlp-main.1055.pdf