CorefInst: Leveraging LLMs for Multilingual Coreference Resolution

Tuğba Pamay Arslan, Emircan Erol, Gülşen Eryiğit


Abstract
Coreference Resolution (CR) is a crucial yet challenging task in natural language understanding, often constrained by task-specific architectures and encoder-based language models that demand extensive training and lack adaptability. This study introduces the first multilingual CR methodology which leverages decoder-only LLMs to handle both overt and zero mentions. The article explores how to model the CR task for LLMs via five different instruction sets using a controlled inference method. The approach is evaluated across three LLMs: Llama 3.1, Gemma 2, and Mistral 0.3. The results indicate that LLMs, when instruction-tuned with a suitable instruction set, can surpass state-of-the-art task-specific architectures. Specifically, our best model, a fully fine-tuned Llama 3.1 for multilingual CR, outperforms the leading multilingual CR model (i.e., Corpipe 24 single stage variant) by 2 percentage points on average across all languages in the CorefUD v1.2 dataset collection.
Anthology ID:
2026.tacl-1.4
Volume:
Transactions of the Association for Computational Linguistics, Volume 14
Month:
Year:
2026
Address:
Cambridge, MA
Venue:
TACL
SIG:
Publisher:
MIT Press
Note:
Pages:
64–80
Language:
URL:
https://aclanthology.org/2026.tacl-1.4/
DOI:
10.1162/tacl.a.593
Bibkey:
Cite (ACL):
Tuğba Pamay Arslan, Emircan Erol, and Gülşen Eryiğit. 2026. CorefInst: Leveraging LLMs for Multilingual Coreference Resolution. Transactions of the Association for Computational Linguistics, 14:64–80.
Cite (Informal):
CorefInst: Leveraging LLMs for Multilingual Coreference Resolution (Pamay Arslan et al., TACL 2026)
Copy Citation:
PDF:
https://aclanthology.org/2026.tacl-1.4.pdf