Dialogue Act-Aided Backchannel Prediction Using Multi-Task Learning

Wencke Liermann, Yo-Han Park, Yong-Seok Choi, Kong Lee


Abstract
Produced in the form of small injections such as “Yeah!” or “Uh-Huh” by listeners in a conversation, supportive verbal feedback (i.e., backchanneling) is essential for natural dialogue. Highlighting its tight relation to speaker intent and utterance type, we propose a multi-task learning approach that learns textual representations for the task of backchannel prediction in tandem with dialogue act classification. We demonstrate the effectiveness of our approach by improving the prediction of specific backchannels like “Yeah” or “Really?” by up to 2.0% in F1. Additionally, whereas previous models relied on well-established methods to extract audio features, we further pre-train the audio encoder in a self-supervised fashion using voice activity projection. This leads to additional gains of 1.4% in weighted F1.
Anthology ID:
2023.findings-emnlp.1006
Volume:
Findings of the Association for Computational Linguistics: EMNLP 2023
Month:
December
Year:
2023
Address:
Singapore
Editors:
Houda Bouamor, Juan Pino, Kalika Bali
Venue:
Findings
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
15073–15079
Language:
URL:
https://aclanthology.org/2023.findings-emnlp.1006
DOI:
10.18653/v1/2023.findings-emnlp.1006
Bibkey:
Cite (ACL):
Wencke Liermann, Yo-Han Park, Yong-Seok Choi, and Kong Lee. 2023. Dialogue Act-Aided Backchannel Prediction Using Multi-Task Learning. In Findings of the Association for Computational Linguistics: EMNLP 2023, pages 15073–15079, Singapore. Association for Computational Linguistics.
Cite (Informal):
Dialogue Act-Aided Backchannel Prediction Using Multi-Task Learning (Liermann et al., Findings 2023)
Copy Citation:
PDF:
https://aclanthology.org/2023.findings-emnlp.1006.pdf