A Human-machine Collaborative Framework for Evaluating Malevolence in Dialogues

Yangjun Zhang, Pengjie Ren, Maarten de Rijke


Abstract
Conversational dialogue systems (CDSs) are hard to evaluate due to the complexity of natural language. Automatic evaluation of dialogues often shows insufficient correlation with human judgements. Human evaluation is reliable but labor-intensive. We introduce a human-machine collaborative framework, HMCEval, that can guarantee reliability of the evaluation outcomes with reduced human effort. HMCEval casts dialogue evaluation as a sample assignment problem, where we need to decide to assign a sample to a human or a machine for evaluation. HMCEval includes a model confidence estimation module to estimate the confidence of the predicted sample assignment, and a human effort estimation module to estimate the human effort should the sample be assigned to human evaluation, as well as a sample assignment execution module that finds the optimum assignment solution based on the estimated confidence and effort. We assess the performance of HMCEval on the task of evaluating malevolence in dialogues. The experimental results show that HMCEval achieves around 99% evaluation accuracy with half of the human effort spared, showing that HMCEval provides reliable evaluation outcomes while reducing human effort by a large amount.
Anthology ID:
2021.acl-long.436
Volume:
Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers)
Month:
August
Year:
2021
Address:
Online
Editors:
Chengqing Zong, Fei Xia, Wenjie Li, Roberto Navigli
Venues:
ACL | IJCNLP
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
5612–5623
Language:
URL:
https://aclanthology.org/2021.acl-long.436
DOI:
10.18653/v1/2021.acl-long.436
Bibkey:
Cite (ACL):
Yangjun Zhang, Pengjie Ren, and Maarten de Rijke. 2021. A Human-machine Collaborative Framework for Evaluating Malevolence in Dialogues. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), pages 5612–5623, Online. Association for Computational Linguistics.
Cite (Informal):
A Human-machine Collaborative Framework for Evaluating Malevolence in Dialogues (Zhang et al., ACL-IJCNLP 2021)
Copy Citation:
PDF:
https://aclanthology.org/2021.acl-long.436.pdf
Video:
 https://aclanthology.org/2021.acl-long.436.mp4
Code
 repozhang/case_hmceval