Using BERT for choosing classifiers in Mandarin

Jani Järnfors, Guanyi Chen, Kees van Deemter, Rint Sybesma


Abstract
Choosing the most suitable classifier in a linguistic context is a well-known problem in the production of Mandarin and many other languages. The present paper proposes a solution based on BERT, compares this solution to previous neural and rule-based models, and argues that the BERT model performs particularly well on those difficult cases where the classifier adds information to the text.
Anthology ID:
2021.inlg-1.17
Volume:
Proceedings of the 14th International Conference on Natural Language Generation
Month:
August
Year:
2021
Address:
Aberdeen, Scotland, UK
Editors:
Anya Belz, Angela Fan, Ehud Reiter, Yaji Sripada
Venue:
INLG
SIG:
SIGGEN
Publisher:
Association for Computational Linguistics
Note:
Pages:
172–176
Language:
URL:
https://aclanthology.org/2021.inlg-1.17
DOI:
10.18653/v1/2021.inlg-1.17
Bibkey:
Cite (ACL):
Jani Järnfors, Guanyi Chen, Kees van Deemter, and Rint Sybesma. 2021. Using BERT for choosing classifiers in Mandarin. In Proceedings of the 14th International Conference on Natural Language Generation, pages 172–176, Aberdeen, Scotland, UK. Association for Computational Linguistics.
Cite (Informal):
Using BERT for choosing classifiers in Mandarin (Järnfors et al., INLG 2021)
Copy Citation:
PDF:
https://aclanthology.org/2021.inlg-1.17.pdf
Data
Chinese Classifier