Katsushi Matsuda
2024
Uncertainty Quantification for In-Context Learning of Large Language Models
Chen Ling
|
Xujiang Zhao
|
Xuchao Zhang
|
Wei Cheng
|
Yanchi Liu
|
Yiyou Sun
|
Mika Oishi
|
Takao Osaki
|
Katsushi Matsuda
|
Jie Ji
|
Guangji Bai
|
Liang Zhao
|
Haifeng Chen
Proceedings of the 2024 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (Volume 1: Long Papers)
In-context learning has emerged as a groundbreaking ability of Large Language Models (LLMs) and revolutionized various fields by providing a few task-relevant demonstrations in the prompt. However, trustworthy issues with LLM’s response, such as hallucination, have also been actively discussed. Existing works have been devoted to quantifying the uncertainty in LLM’s response, but they often overlook the complex nature of LLMs and the uniqueness of in-context learning. In this work, we delve into the predictive uncertainty of LLMs associated with in-context learning, highlighting that such uncertainties may stem from both the provided demonstrations (aleatoric uncertainty) and ambiguities tied to the model’s configurations (epistemic uncertainty). We propose a novel formulation and corresponding estimation method to quantify both types of uncertainties. The proposed method offers an unsupervised way to understand the prediction of in-context learning in a plug-and-play fashion. Extensive experiments are conducted to demonstrate the effectiveness of the decomposition. The code and data are available at: https://github.com/lingchen0331/UQ_ICL.
2023
Open-ended Commonsense Reasoning with Unrestricted Answer Candidates
Chen Ling
|
Xuchao Zhang
|
Xujiang Zhao
|
Yanchi Liu
|
Wei Cheng
|
Mika Oishi
|
Takao Osaki
|
Katsushi Matsuda
|
Haifeng Chen
|
Liang Zhao
Findings of the Association for Computational Linguistics: EMNLP 2023
Open-ended Commonsense Reasoning is defined as solving a commonsense question without providing 1) a short list of answer candidates and 2) a pre-defined answer scope. Conventional ways of formulating the commonsense question into a question-answering form or utilizing external knowledge to learn retrieval-based methods are less applicable in the open-ended setting due to an inherent challenge. Without pre-defining an answer scope or a few candidates, open-ended commonsense reasoning entails predicting answers by searching over an extremely large searching space. Moreover, most questions require implicit multi-hop reasoning, which presents even more challenges to our problem. In this work, we leverage pre-trained language models to iteratively retrieve reasoning paths on the external knowledge base, which does not require task-specific supervision. The reasoning paths can help to identify the most precise answer to the commonsense question. We conduct experiments on two commonsense benchmark datasets. Compared to other approaches, our proposed method achieves better performance both quantitatively and qualitatively.
Search
Co-authors
- Chen Ling 2
- Xuchao Zhang 2
- Xujiang Zhao 2
- Yanchi Liu 2
- Wei Cheng 2
- show all...