GEESE - Generating and Evaluating Explanations for Semantic Entailment: A CALAMITA Challenge

Andrea Zaninello, Bernardo Magnini


Abstract
In the GEESE challenge, we present a pipeline to evaluate generated explanations for the task of Recognizing Textual Entailment (RTE) in Italian. The challenge focuses on evaluating the impact of generated explanations on the predictive performance of language models. Using a dataset enriched with human-written explanations, we employ two large language models (LLMs) to generate and utilize explanations for semantic relationships between sentence pairs. Our methodology assesses the quality of generated explanations by measuring changes in prediction accuracy when explanations are provided. Through reproducible experimentation, we establish benchmarks against various baseline approaches, demonstrating the potential of explanation injection to enhance model interpretability and performance.
Anthology ID:
2024.clicit-1.133
Volume:
Proceedings of the 10th Italian Conference on Computational Linguistics (CLiC-it 2024)
Month:
December
Year:
2024
Address:
Pisa, Italy
Editors:
Felice Dell'Orletta, Alessandro Lenci, Simonetta Montemagni, Rachele Sprugnoli
Venue:
CLiC-it
SIG:
Publisher:
CEUR Workshop Proceedings
Note:
Pages:
1209–1216
Language:
URL:
https://aclanthology.org/2024.clicit-1.133/
DOI:
Bibkey:
Cite (ACL):
Andrea Zaninello and Bernardo Magnini. 2024. GEESE - Generating and Evaluating Explanations for Semantic Entailment: A CALAMITA Challenge. In Proceedings of the 10th Italian Conference on Computational Linguistics (CLiC-it 2024), pages 1209–1216, Pisa, Italy. CEUR Workshop Proceedings.
Cite (Informal):
GEESE - Generating and Evaluating Explanations for Semantic Entailment: A CALAMITA Challenge (Zaninello & Magnini, CLiC-it 2024)
Copy Citation:
PDF:
https://aclanthology.org/2024.clicit-1.133.pdf