Debojeet Chatterjee
2023
A Study on the Efficiency and Generalization of Light Hybrid Retrievers
Man Luo
|
Shashank Jain
|
Anchit Gupta
|
Arash Einolghozati
|
Barlas Oguz
|
Debojeet Chatterjee
|
Xilun Chen
|
Chitta Baral
|
Peyman Heidari
Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers)
Hybrid retrievers can take advantage of both sparse and dense retrievers. Previous hybrid retrievers leverage indexing-heavy dense retrievers. In this work, we study “Is it possible to reduce the indexing memory of hybrid retrievers without sacrificing performance”? Driven by this question, we leverage an indexing-efficient dense retriever (i.e. DrBoost) and introduce a LITE retriever that further reduces the memory of DrBoost. LITE is jointly trained on contrastive learning and knowledge distillation from DrBoost. Then, we integrate BM25, a sparse retriever, with either LITE or DrBoost to form light hybrid retrievers. Our Hybrid-LITE retriever saves 13× memory while maintaining 98.0% performance of the hybrid retriever of BM25 and DPR. In addition, we study the generalization capacity of our light hybrid retrievers on out-of-domain dataset and a set of adversarial attacks datasets. Experiments showcase that light hybrid retrievers achieve better generalization performance than individual sparse and dense retrievers. Nevertheless, our analysis shows that there is a large room to improve the robustness of retrievers, suggesting a new research direction.
Search
Co-authors
- Man Luo 1
- Shashank Jain 1
- Anchit Gupta 1
- Arash Einolghozati 1
- Barlas Oguz 1
- show all...
Venues
- acl1