Embarrassingly Simple Performance Prediction for Abductive Natural Language Inference

Emīls Kadiķis, Vaibhav Srivastav, Roman Klinger


Abstract
The task of natural language inference (NLI), to decide if a hypothesis entails or contradicts a premise, received considerable attention in recent years. All competitive systems build on top of contextualized representations and make use of transformer architectures for learning an NLI model. When somebody is faced with a particular NLI task, they need to select the best model that is available. This is a time-consuming and resource-intense endeavour. To solve this practical problem, we propose a simple method for predicting the performance without actually fine-tuning the model. We do this by testing how well the pre-trained models perform on the aNLI task when just comparing sentence embeddings with cosine similarity to what kind of performance is achieved when training a classifier on top of these embeddings. We show that the accuracy of the cosine similarity approach correlates strongly with the accuracy of the classification approach with a Pearson correlation coefficient of 0.65. Since the similarity is orders of magnitude faster to compute on a given dataset (less than a minute vs. hours), our method can lead to significant time savings in the process of model selection.
Anthology ID:
2022.naacl-main.441
Volume:
Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies
Month:
July
Year:
2022
Address:
Seattle, United States
Venue:
NAACL
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
6031–6037
Language:
URL:
https://aclanthology.org/2022.naacl-main.441
DOI:
10.18653/v1/2022.naacl-main.441
Bibkey:
Cite (ACL):
Emīls Kadiķis, Vaibhav Srivastav, and Roman Klinger. 2022. Embarrassingly Simple Performance Prediction for Abductive Natural Language Inference. In Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pages 6031–6037, Seattle, United States. Association for Computational Linguistics.
Cite (Informal):
Embarrassingly Simple Performance Prediction for Abductive Natural Language Inference (Kadiķis et al., NAACL 2022)
Copy Citation:
PDF:
https://aclanthology.org/2022.naacl-main.441.pdf
Code
 Vaibhavs10/anli-performance-prediction
Data
ART Dataset