Assessing Factual Reliability of Large Language Model Knowledge

Weixuan Wang, Barry Haddow, Alexandra Birch, Wei Peng


Abstract
The factual knowledge of LLMs is typically evaluated using accuracy, yet this metric does not capture the vulnerability of LLMs to hallucination-inducing factors like prompt and context variability. How do we evaluate the capabilities of LLMs to consistently produce factually correct answers? In this paper, we propose MOdel kNowledge relIabiliTy scORe (MONITOR), a novel metric designed to directly measure LLMs’ factual reliability. MONITOR is designed to compute the distance between the probability distributions of a valid output and its counterparts produced by the same LLM probing the same fact using different styles of prompts and contexts. Experiments on a comprehensive range of 12 LLMs demonstrate the effectiveness of MONITOR in evaluating the factual reliability of LLMs while maintaining a low computational overhead. In addition, we release the FKTC (Factual Knowledge Test Corpus) to foster research along this line https://github.com/Vicky-Wil/MONITOR.
Anthology ID:
2024.naacl-long.46
Volume:
Proceedings of the 2024 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (Volume 1: Long Papers)
Month:
June
Year:
2024
Address:
Mexico City, Mexico
Editors:
Kevin Duh, Helena Gomez, Steven Bethard
Venue:
NAACL
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
805–819
Language:
URL:
https://aclanthology.org/2024.naacl-long.46
DOI:
Bibkey:
Cite (ACL):
Weixuan Wang, Barry Haddow, Alexandra Birch, and Wei Peng. 2024. Assessing Factual Reliability of Large Language Model Knowledge. In Proceedings of the 2024 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (Volume 1: Long Papers), pages 805–819, Mexico City, Mexico. Association for Computational Linguistics.
Cite (Informal):
Assessing Factual Reliability of Large Language Model Knowledge (Wang et al., NAACL 2024)
Copy Citation:
PDF:
https://aclanthology.org/2024.naacl-long.46.pdf
Copyright:
 2024.naacl-long.46.copyright.pdf