Effective Contrastive Weighting for Dense Query Expansion

Xiao Wang, Sean MacAvaney, Craig Macdonald, Iadh Ounis


Abstract
Verbatim queries submitted to search engines often do not sufficiently describe the user’s search intent. Pseudo-relevance feedback (PRF) techniques, which modify a query’srepresentation using the top-ranked documents, have been shown to overcome such inadequacies and improve retrieval effectiveness for both lexical methods (e.g., BM25) and dense methods (e.g., ANCE, ColBERT). For instance, the recent ColBERT-PRF approach heuristically chooses new embeddings to add to the query representation using the inverse document frequency (IDF) of the underlying tokens. However, this heuristic potentially ignores the valuable context encoded by the embeddings. In this work, we present a contrastive solution that learns to select the most useful embeddings for expansion. More specifically, a deep language model-based contrastive weighting model, called CWPRF, is trained to learn to discriminate between relevant and non-relevant documents for semantic search. Our experimental results show that our contrastive weighting model can aid to select useful expansion embeddings and outperform various baselines. In particular, CWPRF can improve nDCG@10 by upto to 4.1% compared to an existing PRF approach for ColBERT while maintaining its efficiency.
Anthology ID:
2023.acl-long.710
Volume:
Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)
Month:
July
Year:
2023
Address:
Toronto, Canada
Editors:
Anna Rogers, Jordan Boyd-Graber, Naoaki Okazaki
Venue:
ACL
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
12688–12704
Language:
URL:
https://aclanthology.org/2023.acl-long.710
DOI:
10.18653/v1/2023.acl-long.710
Bibkey:
Cite (ACL):
Xiao Wang, Sean MacAvaney, Craig Macdonald, and Iadh Ounis. 2023. Effective Contrastive Weighting for Dense Query Expansion. In Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 12688–12704, Toronto, Canada. Association for Computational Linguistics.
Cite (Informal):
Effective Contrastive Weighting for Dense Query Expansion (Wang et al., ACL 2023)
Copy Citation:
PDF:
https://aclanthology.org/2023.acl-long.710.pdf
Video:
 https://aclanthology.org/2023.acl-long.710.mp4