CopyT5: Copy Mechanism and Post-Trained T5 for Speech-Aware Dialogue State Tracking System

Cheonyoung Park, Eunji Ha, Yewon Jeong, Chi-young Kim, Haeun Yu, Joo-won Sung


Abstract
In a real-world environment, Dialogue State Tracking (DST) should use speech recognition results to perform tasks. However, most existing DST research has been conducted in text-based environments. This study aims to build a model that efficiently performs Automatic Speech Recognition-based DST. To operate robustly against speech noise, we used CopyT5, which adopted a copy mechanism, and trained the model using augmented data including speech noise. Furthermore, CopyT5 performed post-training using the masked language modeling method with the MultiWOZ dataset in T5 in order to learn the dialogue context better. The copy mechanism also mitigated name entity errors that may occur during DST generation. Experiments confirmed that data augmentation, post-training, and the copy mechanism effectively improve DST performance.
Anthology ID:
2023.dstc-1.11
Volume:
Proceedings of The Eleventh Dialog System Technology Challenge
Month:
September
Year:
2023
Address:
Prague, Czech Republic
Editors:
Yun-Nung Chen, Paul Crook, Michel Galley, Sarik Ghazarian, Chulaka Gunasekara, Raghav Gupta, Behnam Hedayatnia, Satwik Kottur, Seungwhan Moon, Chen Zhang
Venues:
DSTC | WS
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
89–94
Language:
URL:
https://aclanthology.org/2023.dstc-1.11
DOI:
Bibkey:
Cite (ACL):
Cheonyoung Park, Eunji Ha, Yewon Jeong, Chi-young Kim, Haeun Yu, and Joo-won Sung. 2023. CopyT5: Copy Mechanism and Post-Trained T5 for Speech-Aware Dialogue State Tracking System. In Proceedings of The Eleventh Dialog System Technology Challenge, pages 89–94, Prague, Czech Republic. Association for Computational Linguistics.
Cite (Informal):
CopyT5: Copy Mechanism and Post-Trained T5 for Speech-Aware Dialogue State Tracking System (Park et al., DSTC-WS 2023)
Copy Citation:
PDF:
https://aclanthology.org/2023.dstc-1.11.pdf