%0 Conference Proceedings %T Question and Answer Test-Train Overlap in Open-Domain Question Answering Datasets %A Lewis, Patrick %A Stenetorp, Pontus %A Riedel, Sebastian %Y Merlo, Paola %Y Tiedemann, Jorg %Y Tsarfaty, Reut %S Proceedings of the 16th Conference of the European Chapter of the Association for Computational Linguistics: Main Volume %D 2021 %8 April %I Association for Computational Linguistics %C Online %F lewis-etal-2021-question %X Ideally Open-Domain Question Answering models should exhibit a number of competencies, ranging from simply memorizing questions seen at training time, to answering novel question formulations with answers seen during training, to generalizing to completely novel questions with novel answers. However, single aggregated test set scores do not show the full picture of what capabilities models truly have. In this work, we perform a detailed study of the test sets of three popular open-domain benchmark datasets with respect to these competencies. We find that 30% of test-set questions have a near-duplicate paraphrase in their corresponding train sets. In addition, we find that 60-70% of answers in the test sets are also present in the train sets. Using these findings, we evaluate a variety of popular open-domain models to obtain greater insight into what extent they can generalize, and what drives their overall performance. We find that all models perform substantially worse on questions that cannot be memorized from train sets, with a mean absolute performance difference of 61% between repeated and non-repeated data. Finally we show that simple nearest-neighbor models outperform a BART closed-book QA model, further highlighting the role that train set memorization plays in these benchmarks %R 10.18653/v1/2021.eacl-main.86 %U https://aclanthology.org/2021.eacl-main.86 %U https://doi.org/10.18653/v1/2021.eacl-main.86 %P 1000-1008