@inproceedings{baek-etal-2023-knowledge-augmented,
title = "Knowledge-Augmented Language Model Prompting for Zero-Shot Knowledge Graph Question Answering",
author = "Baek, Jinheon and
Aji, Alham Fikri and
Saffari, Amir",
editor = "Hruschka, Estevam and
Mitchell, Tom and
Rahman, Sajjadur and
Mladeni{\'c}, Dunja and
Grobelnik, Marko",
booktitle = "Proceedings of the First Workshop on Matching From Unstructured and Structured Data (MATCHING 2023)",
month = jul,
year = "2023",
address = "Toronto, ON, Canada",
publisher = "Association for Computational Linguistics",
url = "https://aclanthology.org/2023.matching-1.7",
doi = "10.18653/v1/2023.matching-1.7",
pages = "70--98",
abstract = "Large Language Models (LLMs) are capable of performing zero-shot closed-book question answering tasks, based on their internal knowledge stored in parameters during pre-training. However, such internalized knowledge might be insufficient and incorrect, which could lead LLMs to generate factually wrong answers. Furthermore, fine-tuning LLMs to update their knowledge is expensive. To this end, we propose to augment the knowledge directly in the input of LLMs. Specifically, we first retrieve the relevant facts to the input question from the knowledge graph based on semantic similarities between the question and its associated facts. After that, we prepend the retrieved facts to the input question in the form of the prompt, which is then forwarded to LLMs to generate the answer. Our framework, Knowledge-Augmented language model PromptING (KAPING), requires no model training, thus completely zero-shot. We validate the performance of our KAPING framework on the knowledge graph question answering task, that aims to answer the user{'}s question based on facts over a knowledge graph, on which ours outperforms relevant zero-shot baselines by up to 48{\%} in average, across multiple LLMs of various sizes.",
}
<?xml version="1.0" encoding="UTF-8"?>
<modsCollection xmlns="http://www.loc.gov/mods/v3">
<mods ID="baek-etal-2023-knowledge-augmented">
<titleInfo>
<title>Knowledge-Augmented Language Model Prompting for Zero-Shot Knowledge Graph Question Answering</title>
</titleInfo>
<name type="personal">
<namePart type="given">Jinheon</namePart>
<namePart type="family">Baek</namePart>
<role>
<roleTerm authority="marcrelator" type="text">author</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Alham</namePart>
<namePart type="given">Fikri</namePart>
<namePart type="family">Aji</namePart>
<role>
<roleTerm authority="marcrelator" type="text">author</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Amir</namePart>
<namePart type="family">Saffari</namePart>
<role>
<roleTerm authority="marcrelator" type="text">author</roleTerm>
</role>
</name>
<originInfo>
<dateIssued>2023-07</dateIssued>
</originInfo>
<typeOfResource>text</typeOfResource>
<relatedItem type="host">
<titleInfo>
<title>Proceedings of the First Workshop on Matching From Unstructured and Structured Data (MATCHING 2023)</title>
</titleInfo>
<name type="personal">
<namePart type="given">Estevam</namePart>
<namePart type="family">Hruschka</namePart>
<role>
<roleTerm authority="marcrelator" type="text">editor</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Tom</namePart>
<namePart type="family">Mitchell</namePart>
<role>
<roleTerm authority="marcrelator" type="text">editor</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Sajjadur</namePart>
<namePart type="family">Rahman</namePart>
<role>
<roleTerm authority="marcrelator" type="text">editor</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Dunja</namePart>
<namePart type="family">Mladenić</namePart>
<role>
<roleTerm authority="marcrelator" type="text">editor</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Marko</namePart>
<namePart type="family">Grobelnik</namePart>
<role>
<roleTerm authority="marcrelator" type="text">editor</roleTerm>
</role>
</name>
<originInfo>
<publisher>Association for Computational Linguistics</publisher>
<place>
<placeTerm type="text">Toronto, ON, Canada</placeTerm>
</place>
</originInfo>
<genre authority="marcgt">conference publication</genre>
</relatedItem>
<abstract>Large Language Models (LLMs) are capable of performing zero-shot closed-book question answering tasks, based on their internal knowledge stored in parameters during pre-training. However, such internalized knowledge might be insufficient and incorrect, which could lead LLMs to generate factually wrong answers. Furthermore, fine-tuning LLMs to update their knowledge is expensive. To this end, we propose to augment the knowledge directly in the input of LLMs. Specifically, we first retrieve the relevant facts to the input question from the knowledge graph based on semantic similarities between the question and its associated facts. After that, we prepend the retrieved facts to the input question in the form of the prompt, which is then forwarded to LLMs to generate the answer. Our framework, Knowledge-Augmented language model PromptING (KAPING), requires no model training, thus completely zero-shot. We validate the performance of our KAPING framework on the knowledge graph question answering task, that aims to answer the user’s question based on facts over a knowledge graph, on which ours outperforms relevant zero-shot baselines by up to 48% in average, across multiple LLMs of various sizes.</abstract>
<identifier type="citekey">baek-etal-2023-knowledge-augmented</identifier>
<identifier type="doi">10.18653/v1/2023.matching-1.7</identifier>
<location>
<url>https://aclanthology.org/2023.matching-1.7</url>
</location>
<part>
<date>2023-07</date>
<extent unit="page">
<start>70</start>
<end>98</end>
</extent>
</part>
</mods>
</modsCollection>
%0 Conference Proceedings
%T Knowledge-Augmented Language Model Prompting for Zero-Shot Knowledge Graph Question Answering
%A Baek, Jinheon
%A Aji, Alham Fikri
%A Saffari, Amir
%Y Hruschka, Estevam
%Y Mitchell, Tom
%Y Rahman, Sajjadur
%Y Mladenić, Dunja
%Y Grobelnik, Marko
%S Proceedings of the First Workshop on Matching From Unstructured and Structured Data (MATCHING 2023)
%D 2023
%8 July
%I Association for Computational Linguistics
%C Toronto, ON, Canada
%F baek-etal-2023-knowledge-augmented
%X Large Language Models (LLMs) are capable of performing zero-shot closed-book question answering tasks, based on their internal knowledge stored in parameters during pre-training. However, such internalized knowledge might be insufficient and incorrect, which could lead LLMs to generate factually wrong answers. Furthermore, fine-tuning LLMs to update their knowledge is expensive. To this end, we propose to augment the knowledge directly in the input of LLMs. Specifically, we first retrieve the relevant facts to the input question from the knowledge graph based on semantic similarities between the question and its associated facts. After that, we prepend the retrieved facts to the input question in the form of the prompt, which is then forwarded to LLMs to generate the answer. Our framework, Knowledge-Augmented language model PromptING (KAPING), requires no model training, thus completely zero-shot. We validate the performance of our KAPING framework on the knowledge graph question answering task, that aims to answer the user’s question based on facts over a knowledge graph, on which ours outperforms relevant zero-shot baselines by up to 48% in average, across multiple LLMs of various sizes.
%R 10.18653/v1/2023.matching-1.7
%U https://aclanthology.org/2023.matching-1.7
%U https://doi.org/10.18653/v1/2023.matching-1.7
%P 70-98
Markdown (Informal)
[Knowledge-Augmented Language Model Prompting for Zero-Shot Knowledge Graph Question Answering](https://aclanthology.org/2023.matching-1.7) (Baek et al., MATCHING 2023)
ACL