Pre-Training Methods for Question Reranking

Stefano Campese, Ivano Lauriola, Alessandro Moschitti


Abstract
One interesting approach to Question Answering (QA) is to search for semantically similar questions, which have been answered before. This task is different from answer retrieval as it focuses on questions rather than only on the answers, therefore it requires different model training on different data.In this work, we introduce a novel unsupervised pre-training method specialized for retrieving and ranking questions. This leverages (i) knowledge distillation from a basic question retrieval model, and (ii) new pre-training task and objective for learning to rank questions in terms of their relevance with the query. Our experiments show that (i) the proposed technique achieves state-of-the-art performance on QRC and Quora-match datasets, and (ii) the benefit of combining re-ranking and retrieval models.
Anthology ID:
2024.eacl-short.41
Volume:
Proceedings of the 18th Conference of the European Chapter of the Association for Computational Linguistics (Volume 2: Short Papers)
Month:
March
Year:
2024
Address:
St. Julian’s, Malta
Editors:
Yvette Graham, Matthew Purver
Venue:
EACL
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
469–476
Language:
URL:
https://aclanthology.org/2024.eacl-short.41
DOI:
Bibkey:
Cite (ACL):
Stefano Campese, Ivano Lauriola, and Alessandro Moschitti. 2024. Pre-Training Methods for Question Reranking. In Proceedings of the 18th Conference of the European Chapter of the Association for Computational Linguistics (Volume 2: Short Papers), pages 469–476, St. Julian’s, Malta. Association for Computational Linguistics.
Cite (Informal):
Pre-Training Methods for Question Reranking (Campese et al., EACL 2024)
Copy Citation:
PDF:
https://aclanthology.org/2024.eacl-short.41.pdf