Towards a more Robust Evaluation for Conversational Question Answering

Wissam Siblini, Baris Sayil, Yacine Kessaci


Abstract
With the explosion of chatbot applications, Conversational Question Answering (CQA) has generated a lot of interest in recent years. Among proposals, reading comprehension models which take advantage of the conversation history (previous QA) seem to answer better than those which only consider the current question. Nevertheless, we note that the CQA evaluation protocol has a major limitation. In particular, models are allowed, at each turn of the conversation, to access the ground truth answers of the previous turns. Not only does this severely prevent their applications in fully autonomous chatbots, it also leads to unsuspected biases in their behavior. In this paper, we highlight this effect and propose new tools for evaluation and training in order to guard against the noted issues. The new results that we bring come to reinforce methods of the current state of the art.
Anthology ID:
2021.acl-short.130
Volume:
Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 2: Short Papers)
Month:
August
Year:
2021
Address:
Online
Editors:
Chengqing Zong, Fei Xia, Wenjie Li, Roberto Navigli
Venues:
ACL | IJCNLP
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
1028–1034
Language:
URL:
https://aclanthology.org/2021.acl-short.130
DOI:
10.18653/v1/2021.acl-short.130
Bibkey:
Cite (ACL):
Wissam Siblini, Baris Sayil, and Yacine Kessaci. 2021. Towards a more Robust Evaluation for Conversational Question Answering. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 2: Short Papers), pages 1028–1034, Online. Association for Computational Linguistics.
Cite (Informal):
Towards a more Robust Evaluation for Conversational Question Answering (Siblini et al., ACL-IJCNLP 2021)
Copy Citation:
PDF:
https://aclanthology.org/2021.acl-short.130.pdf
Video:
 https://aclanthology.org/2021.acl-short.130.mp4
Data
CoQASQuAD