Do Language Models’ Words Refer?

Matthew Mandelkern, Tal Linzen


Abstract
What do language models (LMs) do with language? They can produce sequences of (mostly) coherent strings closely resembling English. But do those sentences mean something, or are LMs simply babbling in a convincing simulacrum of language use? We address one aspect of this broad question: whether LMs’ words can refer, that is, achieve “word-to-world” connections. There is prima facie reason to think they do not, since LMs do not interact with the world in the way that ordinary language users do. Drawing on the externalist tradition in philosophy of language, we argue that those appearances are misleading: Even if the inputs to LMs are simply strings of text, they are strings of text with natural histories, and that may suffice for LMs’ words to refer.
Anthology ID:
2024.cl-3.12
Volume:
Computational Linguistics, Volume 50, Issue 3 - September 2024
Month:
September
Year:
2024
Address:
Cambridge, MA
Venue:
CL
SIG:
Publisher:
MIT Press
Note:
Pages:
1191–1200
Language:
URL:
https://aclanthology.org/2024.cl-3.12
DOI:
10.1162/coli_a_00522
Bibkey:
Cite (ACL):
Matthew Mandelkern and Tal Linzen. 2024. Do Language Models’ Words Refer?. Computational Linguistics, 50(3):1191–1200.
Cite (Informal):
Do Language Models’ Words Refer? (Mandelkern & Linzen, CL 2024)
Copy Citation:
PDF:
https://aclanthology.org/2024.cl-3.12.pdf