Does Anaphora Resolution Improve LLM Fine-Tuning for Summarisation?

Yi Chun Lo, Ruslan Mitkov


Abstract
This study investigates whether adding anaphora resolution as a preprocessing step before fine-tuning the text summarisation application in LLM can improve the quality of summary output. Two sets of training with the T5-base model and BART-large model using the SAMSum dataset were conducted. One uses the original text and the other uses the text processed by a simplified version of MARS (Mitkov’s Anaphora Resolution System). The experiment reveals that when T5-base model is fine-tuned on the anaphora-resolved inputs, the ROUGE metrics are improved. In contrast, BART-large model only has a slight improvement after fine-tuning under the same conditions, which is not statistically significant. Further analysis of the generated summaries indicates that anaphora resolution is helpful in semantic alignment.
Anthology ID:
2025.r2lm-1.7
Volume:
Proceedings of the First Workshop on Comparative Performance Evaluation: From Rules to Language Models
Month:
September
Year:
2025
Address:
Varna, Bulgaria
Editors:
Alicia Picazo-Izquierdo, Ernesto Luis Estevanell-Valladares, Ruslan Mitkov, Rafael Muñoz Guillena, Raúl García Cerdá
Venues:
R2LM | WS
SIG:
Publisher:
INCOMA Ltd., Shoumen, Bulgaria
Note:
Pages:
59–66
Language:
URL:
https://aclanthology.org/2025.r2lm-1.7/
DOI:
Bibkey:
Cite (ACL):
Yi Chun Lo and Ruslan Mitkov. 2025. Does Anaphora Resolution Improve LLM Fine-Tuning for Summarisation?. In Proceedings of the First Workshop on Comparative Performance Evaluation: From Rules to Language Models, pages 59–66, Varna, Bulgaria. INCOMA Ltd., Shoumen, Bulgaria.
Cite (Informal):
Does Anaphora Resolution Improve LLM Fine-Tuning for Summarisation? (Lo & Mitkov, R2LM 2025)
Copy Citation:
PDF:
https://aclanthology.org/2025.r2lm-1.7.pdf