NeuralQA: A Usable Library for Question Answering (Contextual Query Expansion + BERT) on Large Datasets

Victor Dibia


Abstract
Existing tools for Question Answering (QA) have challenges that limit their use in practice. They can be complex to set up or integrate with existing infrastructure, do not offer configurable interactive interfaces, and do not cover the full set of subtasks that frequently comprise the QA pipeline (query expansion, retrieval, reading, and explanation/sensemaking). To help address these issues, we introduce NeuralQA - a usable library for QA on large datasets. NeuralQA integrates well with existing infrastructure (e.g., ElasticSearch instances and reader models trained with the HuggingFace Transformers API) and offers helpful defaults for QA subtasks. It introduces and implements contextual query expansion (CQE) using a masked language model (MLM) as well as relevant snippets (RelSnip) - a method for condensing large documents into smaller passages that can be speedily processed by a document reader model. Finally, it offers a flexible user interface to support workflows for research explorations (e.g., visualization of gradient-based explanations to support qualitative inspection of model behaviour) and large scale search deployment. Code and documentation for NeuralQA is available as open source on Github.
Anthology ID:
2020.emnlp-demos.3
Volume:
Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing: System Demonstrations
Month:
October
Year:
2020
Address:
Online
Venue:
EMNLP
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
15–22
Language:
URL:
https://aclanthology.org/2020.emnlp-demos.3
DOI:
10.18653/v1/2020.emnlp-demos.3
Bibkey:
Cite (ACL):
Victor Dibia. 2020. NeuralQA: A Usable Library for Question Answering (Contextual Query Expansion + BERT) on Large Datasets. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing: System Demonstrations, pages 15–22, Online. Association for Computational Linguistics.
Cite (Informal):
NeuralQA: A Usable Library for Question Answering (Contextual Query Expansion + BERT) on Large Datasets (Dibia, EMNLP 2020)
Copy Citation:
PDF:
https://aclanthology.org/2020.emnlp-demos.3.pdf
Code
 victordibia/neuralqa