Assessing Monotonicity Reasoning in Dutch through Natural Language Inference

Gijs Wijnholds


Abstract
In this paper we investigate monotonicity reasoning in Dutch, through a novel Natural Language Inference dataset. Monotonicity reasoning shows to be highly challenging for Transformer-based language models in English and here, we corroborate those findings using a parallel Dutch dataset, obtained by translating the Monotonicity Entailment Dataset of Yanaka et al. (2019). After fine-tuning two Dutch language models BERTje and RobBERT on the Dutch NLI dataset SICK-NL, we find that performance severely drops on the monotonicity reasoning dataset, indicating poor generalization capacity of the models. We provide a detailed analysis of the test results by means of the linguistic annotations in the dataset. We find that models struggle with downward entailing contexts, and argue that this is due to a poor understanding of negation. Additionally, we find that the choice of monotonicity context affects model performance on conjunction and disjunction. We hope that this new resource paves the way for further research in generalization of neural reasoning models in Dutch, and contributes to the development of better language technology for Natural Language Inference, specifically for Dutch.
Anthology ID:
2023.findings-eacl.110
Volume:
Findings of the Association for Computational Linguistics: EACL 2023
Month:
May
Year:
2023
Address:
Dubrovnik, Croatia
Editors:
Andreas Vlachos, Isabelle Augenstein
Venue:
Findings
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
1494–1500
Language:
URL:
https://aclanthology.org/2023.findings-eacl.110
DOI:
10.18653/v1/2023.findings-eacl.110
Bibkey:
Cite (ACL):
Gijs Wijnholds. 2023. Assessing Monotonicity Reasoning in Dutch through Natural Language Inference. In Findings of the Association for Computational Linguistics: EACL 2023, pages 1494–1500, Dubrovnik, Croatia. Association for Computational Linguistics.
Cite (Informal):
Assessing Monotonicity Reasoning in Dutch through Natural Language Inference (Wijnholds, Findings 2023)
Copy Citation:
PDF:
https://aclanthology.org/2023.findings-eacl.110.pdf
Dataset:
 2023.findings-eacl.110.dataset.zip
Video:
 https://aclanthology.org/2023.findings-eacl.110.mp4