Gautier Izacard
2023
Task-aware Retrieval with Instructions
Akari Asai
|
Timo Schick
|
Patrick Lewis
|
Xilun Chen
|
Gautier Izacard
|
Sebastian Riedel
|
Hannaneh Hajishirzi
|
Wen-tau Yih
Findings of the Association for Computational Linguistics: ACL 2023
We study the problem of retrieval with instructions, where users provide explicit descriptions of their intent along with their queries to guide a retrieval system. Our solution is a general-purpose task-aware retrieval system, trained using multi-task instruction tuning and can follow human-written instructions to find relevant documents to a given query. We introduce the first large-scale collection of 37 retrieval datasets with instructions, BERRI, and present TART, a single multi-task retrieval system trained on BERRI with instructions that can adapt to a new task without any parameter updates. TART advances the state of the art on two zero-shot retrieval benchmarks, BEIR and LOTTE, outperforming models up to three times larger. We further introduce a new evaluation setup, X2-Retrieval, to better reflect real-world scenarios in which diverse domains and tasks are pooled. TART significantly outperforms competitive baselines in this setup, further highlighting the effectiveness of guiding retrieval with instructions.
2021
Leveraging Passage Retrieval with Generative Models for Open Domain Question Answering
Gautier Izacard
|
Edouard Grave
Proceedings of the 16th Conference of the European Chapter of the Association for Computational Linguistics: Main Volume
Generative models for open domain question answering have proven to be competitive, without resorting to external knowledge. While promising, this approach requires to use models with billions of parameters, which are expensive to train and query. In this paper, we investigate how much these models can benefit from retrieving text passages, potentially containing evidence. We obtain state-of-the-art results on the Natural Questions and TriviaQA open benchmarks. Interestingly, we observe that the performance of this method significantly improves when increasing the number of retrieved passages. This is evidence that sequence-to-sequence models offers a flexible framework to efficiently aggregate and combine evidence from multiple passages.
Search
Co-authors
- Akari Asai 1
- Timo Schick 1
- Patrick Lewis 1
- Xilun Chen 1
- Sebastian Riedel 1
- show all...