Nathan Young
2024
Abstract Meaning Representation-Based Logic-Driven Data Augmentation for Logical Reasoning
Qiming Bao
|
Alex Peng
|
Zhenyun Deng
|
Wanjun Zhong
|
Gael Gendron
|
Timothy Pistotti
|
Neset Tan
|
Nathan Young
|
Yang Chen
|
Yonghua Zhu
|
Paul Denny
|
Michael Witbrock
|
Jiamou Liu
Findings of the Association for Computational Linguistics: ACL 2024
Combining large language models with logical reasoning enhances their capacity to address problems in a robust and reliable manner. Nevertheless, the intricate nature of logical reasoning poses challenges when gathering reliable data from the web to build comprehensive training datasets, subsequently affecting performance on downstream tasks. To address this, we introduce a novel logic-driven data augmentation approach, AMR-LDA. AMR-LDA converts the original text into an Abstract Meaning Representation (AMR) graph, a structured semantic representation that encapsulates the logical structure of the sentence, upon which operations are performed to generate logically modified AMR graphs. The modified AMR graphs are subsequently converted back into text to create augmented data. Notably, our methodology is architecture-agnostic and enhances both generative large language models, such as GPT-3.5 and GPT-4, through prompt augmentation, and discriminative large language models through contrastive learning with logic-driven data augmentation. Empirical evidence underscores the efficacy of our proposed method with improvement in performance across seven downstream tasks, such as reading comprehension requiring logical reasoning, textual entailment, and natural language inference. Furthermore, our method leads on the ReClor leaderboard. The source code and data are publicly available
2022
AbductionRules: Training Transformers to Explain Unexpected Inputs
Nathan Young
|
Qiming Bao
|
Joshua Bensemann
|
Michael Witbrock
Findings of the Association for Computational Linguistics: ACL 2022
Transformers have recently been shown to be capable of reliably performing logical reasoning over facts and rules expressed in natural language, but abductive reasoning - inference to the best explanation of an unexpected observation - has been underexplored despite significant applications to scientific discovery, common-sense reasoning, and model interpretability. This paper presents AbductionRules, a group of natural language datasets designed to train and test generalisable abduction over natural-language knowledge bases. We use these datasets to finetune pretrained Transformers and discuss their performance, finding that our models learned generalisable abductive techniques but also learned to exploit the structure of our data. Finally, we discuss the viability of this approach to abductive reasoning and ways in which it may be improved in future work.
Search
Co-authors
- Qiming Bao 2
- Michael J. Witbrock 2
- Alex Peng 1
- Zhenyun Deng 1
- Wanjun Zhong 1
- show all...