Humans and language models diverge when predicting repeating text

Aditya Vaidya, Javier Turek, Alexander Huth


Abstract
Language models that are trained on the next-word prediction task have been shown to accurately model human behavior in word prediction and reading speed. In contrast with these findings, we present a scenario in which the performance of humans and LMs diverges. We collected a dataset of human next-word predictions for five stimuli that are formed by repeating spans of text. Human and GPT-2 LM predictions are strongly aligned in the first presentation of a text span, but their performance quickly diverges when memory (or in-context learning) begins to play a role. We traced the cause of this divergence to specific attention heads in a middle layer. Adding a power-law recency bias to these attention heads yielded a model that performs much more similarly to humans. We hope that this scenario will spur future work in bringing LMs closer to human behavior.
Anthology ID:
2023.conll-1.5
Volume:
Proceedings of the 27th Conference on Computational Natural Language Learning (CoNLL)
Month:
December
Year:
2023
Address:
Singapore
Editors:
Jing Jiang, David Reitter, Shumin Deng
Venue:
CoNLL
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
58–69
Language:
URL:
https://aclanthology.org/2023.conll-1.5
DOI:
10.18653/v1/2023.conll-1.5
Bibkey:
Cite (ACL):
Aditya Vaidya, Javier Turek, and Alexander Huth. 2023. Humans and language models diverge when predicting repeating text. In Proceedings of the 27th Conference on Computational Natural Language Learning (CoNLL), pages 58–69, Singapore. Association for Computational Linguistics.
Cite (Informal):
Humans and language models diverge when predicting repeating text (Vaidya et al., CoNLL 2023)
Copy Citation:
PDF:
https://aclanthology.org/2023.conll-1.5.pdf