What’s Wrong? Refining Meeting Summaries with LLM Feedback

Frederic Thomas Kirstein, Terry Lima Ruas, Bela Gipp


Abstract
Meeting summarization has become a critical task since digital encounters have become a common practice. Large language models (LLMs) show great potential in summarization, offering enhanced coherence and context understanding compared to traditional methods. However, they still struggle to maintain relevance and avoid hallucination. We introduce a multi-LLM correction approach for meeting summarization using a two-phase process that mimics the human review process: mistake identification and summary refinement. We release QMSum Mistake, a dataset of 200 automatically generated meeting summaries annotated by humans on nine error types, including structural, omission, and irrelevance errors. Our experiments show that these errors can be identified with high accuracy by an LLM. We transform identified mistakes into actionable feedback to improve the quality of a given summary measured by relevance, informativeness, conciseness, and coherence. This post-hoc refinement effectively improves summary quality by leveraging multiple LLMs to validate output quality. Our multi-LLM approach for meeting summarization shows potential for similar complex text generation tasks requiring robustness, action planning, and discussion towards a goal.
Anthology ID:
2025.coling-main.143
Volume:
Proceedings of the 31st International Conference on Computational Linguistics
Month:
January
Year:
2025
Address:
Abu Dhabi, UAE
Editors:
Owen Rambow, Leo Wanner, Marianna Apidianaki, Hend Al-Khalifa, Barbara Di Eugenio, Steven Schockaert
Venue:
COLING
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
2100–2120
Language:
URL:
https://aclanthology.org/2025.coling-main.143/
DOI:
Bibkey:
Cite (ACL):
Frederic Thomas Kirstein, Terry Lima Ruas, and Bela Gipp. 2025. What’s Wrong? Refining Meeting Summaries with LLM Feedback. In Proceedings of the 31st International Conference on Computational Linguistics, pages 2100–2120, Abu Dhabi, UAE. Association for Computational Linguistics.
Cite (Informal):
What’s Wrong? Refining Meeting Summaries with LLM Feedback (Kirstein et al., COLING 2025)
Copy Citation:
PDF:
https://aclanthology.org/2025.coling-main.143.pdf