Investigating the Effect of Natural Language Explanations on Out-of-Distribution Generalization in Few-shot NLI

Yangqiaoyu Zhou, Chenhao Tan


Abstract
Although neural models have shown strong performance in datasets such as SNLI, they lack the ability to generalize out-of-distribution (OOD). In this work, we formulate a few-shot learning setup and examine the effects of natural language explanations on OOD generalization. We leverage the templates in the HANS dataset and construct templated natural language explanations for each template. Although generated explanations show competitive BLEU scores against ground truth explanations, they fail to improve prediction performance. We further show that generated explanations often hallucinate information and miss key elements that indicate the label.
Anthology ID:
2021.insights-1.17
Volume:
Proceedings of the Second Workshop on Insights from Negative Results in NLP
Month:
November
Year:
2021
Address:
Online and Punta Cana, Dominican Republic
Editors:
João Sedoc, Anna Rogers, Anna Rumshisky, Shabnam Tafreshi
Venue:
insights
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
117–124
Language:
URL:
https://aclanthology.org/2021.insights-1.17
DOI:
10.18653/v1/2021.insights-1.17
Bibkey:
Cite (ACL):
Yangqiaoyu Zhou and Chenhao Tan. 2021. Investigating the Effect of Natural Language Explanations on Out-of-Distribution Generalization in Few-shot NLI. In Proceedings of the Second Workshop on Insights from Negative Results in NLP, pages 117–124, Online and Punta Cana, Dominican Republic. Association for Computational Linguistics.
Cite (Informal):
Investigating the Effect of Natural Language Explanations on Out-of-Distribution Generalization in Few-shot NLI (Zhou & Tan, insights 2021)
Copy Citation:
PDF:
https://aclanthology.org/2021.insights-1.17.pdf
Video:
 https://aclanthology.org/2021.insights-1.17.mp4
Code
 chicagohai/hans-explanations
Data
e-SNLI