Mitigating Harms of LLMs via Knowledge Distillation for a Virtual Museum Tour Guide

Ashley Lewis, Michael White


Abstract
LLMs are known to be very powerful, exhibiting both great benefits and great risk. We seek to leverage the benefits, in particular the ability to be fluent, conversational dialogue agents, while minimizing the risks, such as hallucination and toxic content. In this work we use knowledge distillation to create a virtual museum tour guide dialogue agent, employing ChatGPT as a teacher model for a smaller student model, T5-large. We find the T5 model shows competitive performance, significantly reduces instances of hallucination, and shows promise for reducing toxic content.
Anthology ID:
2023.tllm-1.4
Volume:
Proceedings of the 1st Workshop on Taming Large Language Models: Controllability in the era of Interactive Assistants!
Month:
September
Year:
2023
Address:
Prague, Czech Republic
Editors:
Devamanyu Hazarika, Xiangru Robert Tang, Di Jin
Venues:
TLLM | WS
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
31–45
Language:
URL:
https://aclanthology.org/2023.tllm-1.4
DOI:
Bibkey:
Cite (ACL):
Ashley Lewis and Michael White. 2023. Mitigating Harms of LLMs via Knowledge Distillation for a Virtual Museum Tour Guide. In Proceedings of the 1st Workshop on Taming Large Language Models: Controllability in the era of Interactive Assistants!, pages 31–45, Prague, Czech Republic. Association for Computational Linguistics.
Cite (Informal):
Mitigating Harms of LLMs via Knowledge Distillation for a Virtual Museum Tour Guide (Lewis & White, TLLM-WS 2023)
Copy Citation:
PDF:
https://aclanthology.org/2023.tllm-1.4.pdf