FLARE: An Error Analysis Framework for Diagnosing LLM Classification Failures

Keerthana Madhavan, Luiza Antonie, Stacey Scott


Abstract
When Large Language Models return “Inconclusive” in classification tasks, practitioners are left without insight into what went wrong. This diagnostic gap can delay medical decisions, undermine content moderation, and mislead downstream systems. We present FLARE (Failure Location and Reasoning Evaluation), a framework that transforms opaque failures into seven actionable categories. Applied to 5,400 election-misinformation classifications, FLARE reveals a surprising result: Few-Shot prompting—widely considered a best practice—produced 38× more failures than Zero-Shot, with 70.8% due to simple parsing issues. By exposing hidden failure modes, FLARE addresses critical misunderstandings in LLM deployment with implications across domains.
Anthology ID:
2025.ommm-1.4
Volume:
Proceedings of Interdisciplinary Workshop on Observations of Misunderstood, Misguided and Malicious Use of Language Models
Month:
September
Year:
2025
Address:
Varna, Bulgaria
Editors:
Piotr Przybyła, Matthew Shardlow, Clara Colombatto, Nanna Inie
Venues:
OMMM | WS
SIG:
Publisher:
INCOMA Ltd., Shoumen, Bulgaria
Note:
Pages:
40–44
Language:
URL:
https://aclanthology.org/2025.ommm-1.4/
DOI:
Bibkey:
Cite (ACL):
Keerthana Madhavan, Luiza Antonie, and Stacey Scott. 2025. FLARE: An Error Analysis Framework for Diagnosing LLM Classification Failures. In Proceedings of Interdisciplinary Workshop on Observations of Misunderstood, Misguided and Malicious Use of Language Models, pages 40–44, Varna, Bulgaria. INCOMA Ltd., Shoumen, Bulgaria.
Cite (Informal):
FLARE: An Error Analysis Framework for Diagnosing LLM Classification Failures (Madhavan et al., OMMM 2025)
Copy Citation:
PDF:
https://aclanthology.org/2025.ommm-1.4.pdf