Benchmarking Answer Verification Methods for Question Answering-Based Summarization Evaluation Metrics

Daniel Deutsch, Dan Roth


Abstract
Question answering-based summarization evaluation metrics must automatically determine whether the QA model’s prediction is correct or not, a task known as answer verification. In this work, we benchmark the lexical answer verification methods which have been used by current QA-based metrics as well as two more sophisticated text comparison methods, BERTScore and LERC. We find that LERC out-performs the other methods in some settings while remaining statistically indistinguishable from lexical overlap in others. However, our experiments reveal that improved verification performance does not necessarily translate to overall QA-based metric quality: In some scenarios, using a worse verification method — or using none at all — has comparable performance to using the best verification method, a result that we attribute to properties of the datasets.
Anthology ID:
2022.findings-acl.296
Volume:
Findings of the Association for Computational Linguistics: ACL 2022
Month:
May
Year:
2022
Address:
Dublin, Ireland
Venues:
ACL | Findings
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
3759–3765
Language:
URL:
https://aclanthology.org/2022.findings-acl.296
DOI:
10.18653/v1/2022.findings-acl.296
Bibkey:
Cite (ACL):
Daniel Deutsch and Dan Roth. 2022. Benchmarking Answer Verification Methods for Question Answering-Based Summarization Evaluation Metrics. In Findings of the Association for Computational Linguistics: ACL 2022, pages 3759–3765, Dublin, Ireland. Association for Computational Linguistics.
Cite (Informal):
Benchmarking Answer Verification Methods for Question Answering-Based Summarization Evaluation Metrics (Deutsch & Roth, Findings 2022)
Copy Citation:
PDF:
https://aclanthology.org/2022.findings-acl.296.pdf