End to End Multilingual Coreference Resolution for Indian Languages

Sobha Lalitha Devi, Vijay Sundar Ram, Pattabhi RK Rao


Abstract
This paper describes an approach on an end to end model for Multilingual Coreference Resolution (CR) for low resource languages such as Tamil, Malayalam and Hindi. We have done fine tune the XLM-Roberta large model on multilingual training dataset using specific languages with linguistic features and without linguistic features. XLM-R with linguistic features achieves better results than the baseline system. This shows that giving the linguistic knowledge enriches the system performance. The performance of the system is comparable with the state of the art systems.
Anthology ID:
2024.icon-1.29
Volume:
Proceedings of the 21st International Conference on Natural Language Processing (ICON)
Month:
December
Year:
2024
Address:
AU-KBC Research Centre, Chennai, India
Editors:
Sobha Lalitha Devi, Karunesh Arora
Venue:
ICON
SIG:
Publisher:
NLP Association of India (NLPAI)
Note:
Pages:
256–259
Language:
URL:
https://aclanthology.org/2024.icon-1.29/
DOI:
Bibkey:
Cite (ACL):
Sobha Lalitha Devi, Vijay Sundar Ram, and Pattabhi RK Rao. 2024. End to End Multilingual Coreference Resolution for Indian Languages. In Proceedings of the 21st International Conference on Natural Language Processing (ICON), pages 256–259, AU-KBC Research Centre, Chennai, India. NLP Association of India (NLPAI).
Cite (Informal):
End to End Multilingual Coreference Resolution for Indian Languages (Lalitha Devi et al., ICON 2024)
Copy Citation:
PDF:
https://aclanthology.org/2024.icon-1.29.pdf