AutoReply: Detecting Nonsense in Dialogue with Discriminative Replies

Weiyan Shi, Emily Dinan, Adi Renduchintala, Daniel Fried, Athul Jacob, Zhou Yu, Mike Lewis


Abstract
We show that dialogue models can detect errors in their own messages, by calculating the likelihood of replies that are indicative of poor messages. For example, if an agent believes its partner is likely to respond “I don’t understand” to a candidate message, that message may not make sense, so an alternative message should be chosen. We evaluate our approach on a dataset from the game Diplomacy, which contains long dialogues richly grounded in the game state, on which existing models make many errors. We first show that hand-crafted replies can be effective for the task of detecting nonsense in applications as complex as Diplomacy. We then design AutoReply, an algorithm to search for such discriminative replies automatically, given a small number of annotated dialogue examples. We find that AutoReply-generated replies outperform handcrafted replies and perform on par with supervised learning approaches.
Anthology ID:
2023.findings-emnlp.23
Volume:
Findings of the Association for Computational Linguistics: EMNLP 2023
Month:
December
Year:
2023
Address:
Singapore
Editors:
Houda Bouamor, Juan Pino, Kalika Bali
Venue:
Findings
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
294–309
Language:
URL:
https://aclanthology.org/2023.findings-emnlp.23
DOI:
10.18653/v1/2023.findings-emnlp.23
Bibkey:
Cite (ACL):
Weiyan Shi, Emily Dinan, Adi Renduchintala, Daniel Fried, Athul Jacob, Zhou Yu, and Mike Lewis. 2023. AutoReply: Detecting Nonsense in Dialogue with Discriminative Replies. In Findings of the Association for Computational Linguistics: EMNLP 2023, pages 294–309, Singapore. Association for Computational Linguistics.
Cite (Informal):
AutoReply: Detecting Nonsense in Dialogue with Discriminative Replies (Shi et al., Findings 2023)
Copy Citation:
PDF:
https://aclanthology.org/2023.findings-emnlp.23.pdf