Investigating the Impact of ASR Errors on Spoken Implicit Discourse Relation Recognition

Linh The Nguyen, Dat Quoc Nguyen


Abstract
We present an empirical study investigating the influence of automatic speech recognition (ASR) errors on the spoken implicit discourse relation recognition (IDRR) task. We construct a spoken dataset for this task based on the Penn Discourse Treebank 2.0. On this dataset, we conduct “Cascaded” experiments employing state-of-the-art ASR and text-based IDRR models and find that the ASR errors significantly decrease the IDRR performance. In addition, the “Cascaded” approach does remarkably better than an “End-to-End” one that directly predicts a relation label for each input argument speech pair.
Anthology ID:
2022.tu-1.5
Volume:
Proceedings of the First Workshop On Transcript Understanding
Month:
Oct
Year:
2022
Address:
Gyeongju, South Korea
Editors:
Franck Dernoncourt, Thien Huu Nguyen, Viet Dac Lai, Amir Pouran Ben Veyseh, Trung H. Bui, David Seunghyun Yoon
Venue:
TU
SIG:
Publisher:
International Conference on Computational Linguistics
Note:
Pages:
34–39
Language:
URL:
https://aclanthology.org/2022.tu-1.5
DOI:
Bibkey:
Cite (ACL):
Linh The Nguyen and Dat Quoc Nguyen. 2022. Investigating the Impact of ASR Errors on Spoken Implicit Discourse Relation Recognition. In Proceedings of the First Workshop On Transcript Understanding, pages 34–39, Gyeongju, South Korea. International Conference on Computational Linguistics.
Cite (Informal):
Investigating the Impact of ASR Errors on Spoken Implicit Discourse Relation Recognition (Nguyen & Nguyen, TU 2022)
Copy Citation:
PDF:
https://aclanthology.org/2022.tu-1.5.pdf