Cognitive Biases, Task Complexity, and Result Intepretability in Large Language Models

Mario Mina, Valle Ruiz-Fernández, Júlia Falcão, Luis Vasquez-Reina, Aitor Gonzalez-Agirre


Abstract
In humans, cognitive biases are systematic deviations from rationality in judgment that simplify complex decisions. They typically manifest as a consequence of learned behaviors or limitations on information processing capabilities. Recent work has shown that these biases can percolate through training data and ultimately be learned by language models. We examine different groups of models, factoring in model size and type (base or instructed) for four kinds of cognitive bias: primacy, recency, common token, and majority class bias. We evaluate the performance of each model for each type of bias in different settings using simple and complex variants of datasets. Our results show that some biases have much stronger effects than others, and that task complexity plays a part in eliciting stronger effects for some of these biases as measured by effect size. We show that some cognitive biases such as common token and majority class bias are not straightforward to evaluate, and that, contrary to some of the previous literature, some effects that have been previously classified as common token bias in the literature are actually due to primacy and recency bias.
Anthology ID:
2025.coling-main.120
Volume:
Proceedings of the 31st International Conference on Computational Linguistics
Month:
January
Year:
2025
Address:
Abu Dhabi, UAE
Editors:
Owen Rambow, Leo Wanner, Marianna Apidianaki, Hend Al-Khalifa, Barbara Di Eugenio, Steven Schockaert
Venue:
COLING
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
1767–1784
Language:
URL:
https://aclanthology.org/2025.coling-main.120/
DOI:
Bibkey:
Cite (ACL):
Mario Mina, Valle Ruiz-Fernández, Júlia Falcão, Luis Vasquez-Reina, and Aitor Gonzalez-Agirre. 2025. Cognitive Biases, Task Complexity, and Result Intepretability in Large Language Models. In Proceedings of the 31st International Conference on Computational Linguistics, pages 1767–1784, Abu Dhabi, UAE. Association for Computational Linguistics.
Cite (Informal):
Cognitive Biases, Task Complexity, and Result Intepretability in Large Language Models (Mina et al., COLING 2025)
Copy Citation:
PDF:
https://aclanthology.org/2025.coling-main.120.pdf