@inproceedings{iarosh-etal-2025-reducing,
title = "On Reducing Factual Hallucinations in Graph-to-Text Generation Using Large Language Models",
author = "Iarosh, Dmitrii and
Panchenko, Alexander and
Salnikov, Mikhail",
editor = "Gesese, Genet Asefa and
Sack, Harald and
Paulheim, Heiko and
Merono-Penuela, Albert and
Chen, Lihu",
booktitle = "Proceedings of the Workshop on Generative AI and Knowledge Graphs (GenAIK)",
month = jan,
year = "2025",
address = "Abu Dhabi, UAE",
publisher = "International Committee on Computational Linguistics",
url = "https://aclanthology.org/2025.genaik-1.5/",
pages = "43--53",
abstract = "Recent work in Graph-to-Text generation has achieved impressive results, but it still suffers from hallucinations in some cases, despite extensive pretraining stages and various methods for working with graph data. While the commonly used metrics for evaluating the quality of Graph-to-Text models show almost perfect results, it makes it challenging to compare different approaches. This paper demonstrates the challenges of recent Graph-to-Text systems in terms of hallucinations and proposes a simple yet effective approach to using a general LLM, which has shown state-of-the-art results and reduced the number of factual hallucinations. We provide step-by-step instructions on how to develop prompts for language models and a detailed analysis of potential factual errors in the generated text."
}
<?xml version="1.0" encoding="UTF-8"?>
<modsCollection xmlns="http://www.loc.gov/mods/v3">
<mods ID="iarosh-etal-2025-reducing">
<titleInfo>
<title>On Reducing Factual Hallucinations in Graph-to-Text Generation Using Large Language Models</title>
</titleInfo>
<name type="personal">
<namePart type="given">Dmitrii</namePart>
<namePart type="family">Iarosh</namePart>
<role>
<roleTerm authority="marcrelator" type="text">author</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Alexander</namePart>
<namePart type="family">Panchenko</namePart>
<role>
<roleTerm authority="marcrelator" type="text">author</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Mikhail</namePart>
<namePart type="family">Salnikov</namePart>
<role>
<roleTerm authority="marcrelator" type="text">author</roleTerm>
</role>
</name>
<originInfo>
<dateIssued>2025-01</dateIssued>
</originInfo>
<typeOfResource>text</typeOfResource>
<relatedItem type="host">
<titleInfo>
<title>Proceedings of the Workshop on Generative AI and Knowledge Graphs (GenAIK)</title>
</titleInfo>
<name type="personal">
<namePart type="given">Genet</namePart>
<namePart type="given">Asefa</namePart>
<namePart type="family">Gesese</namePart>
<role>
<roleTerm authority="marcrelator" type="text">editor</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Harald</namePart>
<namePart type="family">Sack</namePart>
<role>
<roleTerm authority="marcrelator" type="text">editor</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Heiko</namePart>
<namePart type="family">Paulheim</namePart>
<role>
<roleTerm authority="marcrelator" type="text">editor</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Albert</namePart>
<namePart type="family">Merono-Penuela</namePart>
<role>
<roleTerm authority="marcrelator" type="text">editor</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Lihu</namePart>
<namePart type="family">Chen</namePart>
<role>
<roleTerm authority="marcrelator" type="text">editor</roleTerm>
</role>
</name>
<originInfo>
<publisher>International Committee on Computational Linguistics</publisher>
<place>
<placeTerm type="text">Abu Dhabi, UAE</placeTerm>
</place>
</originInfo>
<genre authority="marcgt">conference publication</genre>
</relatedItem>
<abstract>Recent work in Graph-to-Text generation has achieved impressive results, but it still suffers from hallucinations in some cases, despite extensive pretraining stages and various methods for working with graph data. While the commonly used metrics for evaluating the quality of Graph-to-Text models show almost perfect results, it makes it challenging to compare different approaches. This paper demonstrates the challenges of recent Graph-to-Text systems in terms of hallucinations and proposes a simple yet effective approach to using a general LLM, which has shown state-of-the-art results and reduced the number of factual hallucinations. We provide step-by-step instructions on how to develop prompts for language models and a detailed analysis of potential factual errors in the generated text.</abstract>
<identifier type="citekey">iarosh-etal-2025-reducing</identifier>
<location>
<url>https://aclanthology.org/2025.genaik-1.5/</url>
</location>
<part>
<date>2025-01</date>
<extent unit="page">
<start>43</start>
<end>53</end>
</extent>
</part>
</mods>
</modsCollection>
%0 Conference Proceedings
%T On Reducing Factual Hallucinations in Graph-to-Text Generation Using Large Language Models
%A Iarosh, Dmitrii
%A Panchenko, Alexander
%A Salnikov, Mikhail
%Y Gesese, Genet Asefa
%Y Sack, Harald
%Y Paulheim, Heiko
%Y Merono-Penuela, Albert
%Y Chen, Lihu
%S Proceedings of the Workshop on Generative AI and Knowledge Graphs (GenAIK)
%D 2025
%8 January
%I International Committee on Computational Linguistics
%C Abu Dhabi, UAE
%F iarosh-etal-2025-reducing
%X Recent work in Graph-to-Text generation has achieved impressive results, but it still suffers from hallucinations in some cases, despite extensive pretraining stages and various methods for working with graph data. While the commonly used metrics for evaluating the quality of Graph-to-Text models show almost perfect results, it makes it challenging to compare different approaches. This paper demonstrates the challenges of recent Graph-to-Text systems in terms of hallucinations and proposes a simple yet effective approach to using a general LLM, which has shown state-of-the-art results and reduced the number of factual hallucinations. We provide step-by-step instructions on how to develop prompts for language models and a detailed analysis of potential factual errors in the generated text.
%U https://aclanthology.org/2025.genaik-1.5/
%P 43-53
Markdown (Informal)
[On Reducing Factual Hallucinations in Graph-to-Text Generation Using Large Language Models](https://aclanthology.org/2025.genaik-1.5/) (Iarosh et al., GenAIK 2025)
ACL