Antonio Mallia
2026
Statistical Foundations of DIME: Risk Estimation for Practical Index Selection
Giulio D'Erasmo | Cesare Campagnano | Antonio Mallia | Pierpaolo Brutti | Nicola Tonellotto | Fabrizio Silvestri
Proceedings of the 19th Conference of the European Chapter of the Association for Computational Linguistics (Volume 2: Short Papers)
Giulio D'Erasmo | Cesare Campagnano | Antonio Mallia | Pierpaolo Brutti | Nicola Tonellotto | Fabrizio Silvestri
Proceedings of the 19th Conference of the European Chapter of the Association for Computational Linguistics (Volume 2: Short Papers)
High-dimensional dense embeddings have become central to modern Information Retrieval, but many dimensions are noisy or redundant. Recently proposed DIME (Dimension IMportance Estimation), provides query-dependent scores to identify informative components of embeddings. DIME relies on a costly grid search to select a priori a dimensionality for all the query corpus’s embeddings. Our work provides a statistically grounded criterion that directly identifies the optimal set of dimensions for each query at inference time. Experiments confirm that this approach improves retrieval effectiveness and reduces embedding size by an average 50% of across different models and datasets at inference time.
2022
Accelerating Learned Sparse Indexes Via Term Impact Decomposition
Joel Mackenzie | Antonio Mallia | Alistair Moffat | Matthias Petri
Findings of the Association for Computational Linguistics: EMNLP 2022
Joel Mackenzie | Antonio Mallia | Alistair Moffat | Matthias Petri
Findings of the Association for Computational Linguistics: EMNLP 2022
Novel inverted index-based learned sparse ranking models provide more effective, but less efficient, retrieval performance compared to traditional ranking models like BM25. In this paper, we introduce a technique we call postings clipping to improve the query efficiency of learned representations. Our technique amplifies the benefit of dynamic pruning query processing techniques by accounting for changes in term importance distributions of learned ranking models. The new clipping mechanism accelerates top-k retrieval by up to 9.6X without any loss in effectiveness.