Internet-Augmented Dialogue Generation

Mojtaba Komeili, Kurt Shuster, Jason Weston


Abstract
The largest store of continually updating knowledge on our planet can be accessed via internet search. In this work we study giving access to this information to conversational agents. Large language models, even though they store an impressive amount of knowledge within their weights, are known to hallucinate facts when generating dialogue (Shuster et al., 2021); moreover, those facts are frozen in time at the point of model training. In contrast, we propose an approach that learns to generate an internet search query based on the context, and then conditions on the search results to finally generate a response, a method that can employ up-to-the-minute relevant information. We train and evaluate such models on a newly collected dataset of human-human conversations whereby one of the speakers is given access to internet search during knowledgedriven discussions in order to ground their responses. We find that search-query based access of the internet in conversation provides superior performance compared to existing approaches that either use no augmentation or FAISS-based retrieval (Lewis et al., 2020b).
Anthology ID:
2022.acl-long.579
Volume:
Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)
Month:
May
Year:
2022
Address:
Dublin, Ireland
Editors:
Smaranda Muresan, Preslav Nakov, Aline Villavicencio
Venue:
ACL
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
8460–8478
Language:
URL:
https://aclanthology.org/2022.acl-long.579
DOI:
10.18653/v1/2022.acl-long.579
Bibkey:
Cite (ACL):
Mojtaba Komeili, Kurt Shuster, and Jason Weston. 2022. Internet-Augmented Dialogue Generation. In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 8460–8478, Dublin, Ireland. Association for Computational Linguistics.
Cite (Informal):
Internet-Augmented Dialogue Generation (Komeili et al., ACL 2022)
Copy Citation:
PDF:
https://aclanthology.org/2022.acl-long.579.pdf
Video:
 https://aclanthology.org/2022.acl-long.579.mp4
Data
Topical-ChatWizard of Wikipedia