Mind the Biases: Quantifying Cognitive Biases in Language Model Prompting

Ruixi Lin, Hwee Tou Ng


Abstract
We advocate the importance of exposing uncertainty on results of language model prompting which display bias modes resembling cognitive biases, and propose to help users grasp the level of uncertainty via simple quantifying metrics. Cognitive biases in the human decision making process can lead to flawed responses when we are under uncertainty. Not surprisingly, we have seen biases in language models resembling cognitive biases as a result of training on biased textual data, raising dangers in downstream tasks that are centered around people’s lives if users trust their results too much. In this work, we reveal two bias modes leveraging cognitive biases when we prompt BERT, accompanied by two bias metrics. On a drug-drug interaction extraction task, our bias measurements reveal an error pattern similar to the availability bias when the labels for training prompts are imbalanced, and show that a toning-down transformation of the drug-drug description in a prompt can elicit a bias similar to the framing effect, warning users to distrust when prompting language models for answers.
Anthology ID:
2023.findings-acl.324
Volume:
Findings of the Association for Computational Linguistics: ACL 2023
Month:
July
Year:
2023
Address:
Toronto, Canada
Editors:
Anna Rogers, Jordan Boyd-Graber, Naoaki Okazaki
Venue:
Findings
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
5269–5281
Language:
URL:
https://aclanthology.org/2023.findings-acl.324
DOI:
10.18653/v1/2023.findings-acl.324
Bibkey:
Cite (ACL):
Ruixi Lin and Hwee Tou Ng. 2023. Mind the Biases: Quantifying Cognitive Biases in Language Model Prompting. In Findings of the Association for Computational Linguistics: ACL 2023, pages 5269–5281, Toronto, Canada. Association for Computational Linguistics.
Cite (Informal):
Mind the Biases: Quantifying Cognitive Biases in Language Model Prompting (Lin & Ng, Findings 2023)
Copy Citation:
PDF:
https://aclanthology.org/2023.findings-acl.324.pdf