Zhixin Yin
2024
LawBench: Benchmarking Legal Knowledge of Large Language Models
Zhiwei Fei
|
Xiaoyu Shen
|
Dawei Zhu
|
Fengzhe Zhou
|
Zhuo Han
|
Alan Huang
|
Songyang Zhang
|
Kai Chen
|
Zhixin Yin
|
Zongwen Shen
|
Jidong Ge
|
Vincent Ng
Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing
We present LawBench, the first evaluation benchmark composed of 20 tasks aimed to assess the ability of Large Language Models (LLMs) to perform Chinese legal-related tasks. LawBench is meticulously crafted to enable precise assessment of LLMs’ legal capabilities from three cognitive levels that correspond to the widely accepted Bloom’s cognitive taxonomy. Using LawBench, we present a comprehensive evaluation of 21 popular LLMs and the first comparative analysis of the empirical results in order to reveal their relative strengths and weaknesses. All data, model predictions and evaluation code are accessible from https://github.com/open-compass/LawBench.
Search
Co-authors
- Zhiwei Fei 1
- Xiaoyu Shen 1
- Dawei Zhu 1
- Fengzhe Zhou 1
- Zhuo Han 1
- show all...