@inproceedings{noble-etal-2025-mood,
title = "In the Mood for Inference: Logic-Based Natural Language Inference with Large Language Models",
author = "Noble, Bill and
Blanck, Rasmus and
Wijnholds, Gijs",
editor = "Abzianidze, Lasha and
de Paiva, Valeria",
booktitle = "Proceedings of the 5th Workshop on Natural Logic Meets Machine Learning (NALOMA)",
month = aug,
year = "2025",
address = "Bochum, Germany",
publisher = "Association for Computational Linguistics",
url = "https://aclanthology.org/2025.naloma-1.4/",
pages = "33--47",
ISBN = "979-8-89176-287-9",
abstract = "In this paper we explore challenging Natural Language Inference datasets with a logic-based approach and Large Language Models (LLMs), in order to assess the validity of this hybrid strategy. We report on an experiment which combines an LLM meta-prompting strategy, eliciting logical representations, and Prover9, a first-order logic theorem prover. In addition, we experiment with the inclusion of (logical) world knowledge. Our findings suggest that (i) broad performance is sometimes on par, (ii) formula generation is rather brittle, and (iii) world knowledge aids performance relative to data annotation. We argue that these results explicate the weaknesses of both approaches. As such, we consider this study a source of inspiration for future work in the field of neuro-symbolic reasoning."
}
<?xml version="1.0" encoding="UTF-8"?>
<modsCollection xmlns="http://www.loc.gov/mods/v3">
<mods ID="noble-etal-2025-mood">
<titleInfo>
<title>In the Mood for Inference: Logic-Based Natural Language Inference with Large Language Models</title>
</titleInfo>
<name type="personal">
<namePart type="given">Bill</namePart>
<namePart type="family">Noble</namePart>
<role>
<roleTerm authority="marcrelator" type="text">author</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Rasmus</namePart>
<namePart type="family">Blanck</namePart>
<role>
<roleTerm authority="marcrelator" type="text">author</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Gijs</namePart>
<namePart type="family">Wijnholds</namePart>
<role>
<roleTerm authority="marcrelator" type="text">author</roleTerm>
</role>
</name>
<originInfo>
<dateIssued>2025-08</dateIssued>
</originInfo>
<typeOfResource>text</typeOfResource>
<relatedItem type="host">
<titleInfo>
<title>Proceedings of the 5th Workshop on Natural Logic Meets Machine Learning (NALOMA)</title>
</titleInfo>
<name type="personal">
<namePart type="given">Lasha</namePart>
<namePart type="family">Abzianidze</namePart>
<role>
<roleTerm authority="marcrelator" type="text">editor</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Valeria</namePart>
<namePart type="family">de Paiva</namePart>
<role>
<roleTerm authority="marcrelator" type="text">editor</roleTerm>
</role>
</name>
<originInfo>
<publisher>Association for Computational Linguistics</publisher>
<place>
<placeTerm type="text">Bochum, Germany</placeTerm>
</place>
</originInfo>
<genre authority="marcgt">conference publication</genre>
<identifier type="isbn">979-8-89176-287-9</identifier>
</relatedItem>
<abstract>In this paper we explore challenging Natural Language Inference datasets with a logic-based approach and Large Language Models (LLMs), in order to assess the validity of this hybrid strategy. We report on an experiment which combines an LLM meta-prompting strategy, eliciting logical representations, and Prover9, a first-order logic theorem prover. In addition, we experiment with the inclusion of (logical) world knowledge. Our findings suggest that (i) broad performance is sometimes on par, (ii) formula generation is rather brittle, and (iii) world knowledge aids performance relative to data annotation. We argue that these results explicate the weaknesses of both approaches. As such, we consider this study a source of inspiration for future work in the field of neuro-symbolic reasoning.</abstract>
<identifier type="citekey">noble-etal-2025-mood</identifier>
<location>
<url>https://aclanthology.org/2025.naloma-1.4/</url>
</location>
<part>
<date>2025-08</date>
<extent unit="page">
<start>33</start>
<end>47</end>
</extent>
</part>
</mods>
</modsCollection>
%0 Conference Proceedings
%T In the Mood for Inference: Logic-Based Natural Language Inference with Large Language Models
%A Noble, Bill
%A Blanck, Rasmus
%A Wijnholds, Gijs
%Y Abzianidze, Lasha
%Y de Paiva, Valeria
%S Proceedings of the 5th Workshop on Natural Logic Meets Machine Learning (NALOMA)
%D 2025
%8 August
%I Association for Computational Linguistics
%C Bochum, Germany
%@ 979-8-89176-287-9
%F noble-etal-2025-mood
%X In this paper we explore challenging Natural Language Inference datasets with a logic-based approach and Large Language Models (LLMs), in order to assess the validity of this hybrid strategy. We report on an experiment which combines an LLM meta-prompting strategy, eliciting logical representations, and Prover9, a first-order logic theorem prover. In addition, we experiment with the inclusion of (logical) world knowledge. Our findings suggest that (i) broad performance is sometimes on par, (ii) formula generation is rather brittle, and (iii) world knowledge aids performance relative to data annotation. We argue that these results explicate the weaknesses of both approaches. As such, we consider this study a source of inspiration for future work in the field of neuro-symbolic reasoning.
%U https://aclanthology.org/2025.naloma-1.4/
%P 33-47
Markdown (Informal)
[In the Mood for Inference: Logic-Based Natural Language Inference with Large Language Models](https://aclanthology.org/2025.naloma-1.4/) (Noble et al., NALOMA 2025)
ACL