MoQA: Benchmarking Multi-Type Open-Domain Question Answering

Howard Yen, Tianyu Gao, Jinhyuk Lee, Danqi Chen


Abstract
Previous research on open-domain question answering (QA) mainly focuses on questions with short answers. However, information-seeking QA often requires various formats of answers depending on the nature of the questions, e.g., why/how questions typically require a long answer. In this paper, we present MoQA, a benchmark for open-domain QA that requires building one system that can provide short, medium, long, and yes/no answers to different questions accordingly. MoQA builds upon Natural Questions with multiple types of questions and additional crowdsourcing efforts to ensure high query quality. We adapt state-of-the-art models, and reveal unique findings in multi-type open-domain QA: (1) For retriever-reader models, training one retriever on all types achieves the overall best performance, but it is challenging to train one reader model to output answers of different formats, or to train a question classifier to distinguish between types; (2) An end-to-end closed-book QA model trained on multiple types struggles with the task across the board; (3) State-of-the-art large language models such as the largest GPT-3 models (Brown et al., 2020; Ouyang et al., 2022) also lag behind open-book QA models. Our benchmark and analysis call for more effort into building versatile open-domain QA models in the future.
Anthology ID:
2023.dialdoc-1.2
Volume:
Proceedings of the Third DialDoc Workshop on Document-grounded Dialogue and Conversational Question Answering
Month:
July
Year:
2023
Address:
Toronto, Canada
Editors:
Smaranda Muresan, Vivian Chen, Kennington Casey, Vandyke David, Dethlefs Nina, Inoue Koji, Ekstedt Erik, Ultes Stefan
Venue:
dialdoc
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
8–29
Language:
URL:
https://aclanthology.org/2023.dialdoc-1.2
DOI:
10.18653/v1/2023.dialdoc-1.2
Bibkey:
Cite (ACL):
Howard Yen, Tianyu Gao, Jinhyuk Lee, and Danqi Chen. 2023. MoQA: Benchmarking Multi-Type Open-Domain Question Answering. In Proceedings of the Third DialDoc Workshop on Document-grounded Dialogue and Conversational Question Answering, pages 8–29, Toronto, Canada. Association for Computational Linguistics.
Cite (Informal):
MoQA: Benchmarking Multi-Type Open-Domain Question Answering (Yen et al., dialdoc 2023)
Copy Citation:
PDF:
https://aclanthology.org/2023.dialdoc-1.2.pdf