Subeen Pang
2023
Retrieval-based Evaluation for LLMs: A Case Study in Korean Legal QA
Cheol Ryu
|
Seolhwa Lee
|
Subeen Pang
|
Chanyeol Choi
|
Hojun Choi
|
Myeonggee Min
|
Jy-Yong Sohn
Proceedings of the Natural Legal Language Processing Workshop 2023
While large language models (LLMs) have demonstrated significant capabilities in text generation, their utilization in areas requiring domain-specific expertise, such as law, must be approached cautiously. This caution is warranted due to the inherent challenges associated with LLM-generated texts, including the potential presence of factual errors. Motivated by this issue, we propose Eval-RAG, a new evaluation method for LLM-generated texts. Unlike existing methods, Eval-RAG evaluates the validity of generated texts based on the related document that are collected by the retriever. In other words, Eval-RAG adopts the idea of retrieval augmented generation (RAG) for the purpose of evaluation. Our experimental results on Korean Legal Question-Answering (QA) tasks show that conventional LLM-based evaluation methods can be better aligned with Lawyers’ evaluations, by combining with Eval-RAG. In addition, our qualitative analysis show that Eval-RAG successfully finds the factual errors in LLM-generated texts, while existing evaluation methods cannot.
Search
Co-authors
- Cheol Ryu 1
- Seolhwa Lee 1
- Chanyeol Choi 1
- Hojun Choi 1
- Myeonggee Min 1
- show all...