Adrian Pol
2024
Measuring the Groundedness of Legal Question-Answering Systems
Dietrich Trautmann
|
Natalia Ostapuk
|
Quentin Grail
|
Adrian Pol
|
Guglielmo Bonifazi
|
Shang Gao
|
Martin Gajek
Proceedings of the Natural Legal Language Processing Workshop 2024
In high-stakes domains like legal question-answering, the accuracy and trustworthiness of generative AI systems are of paramount importance. This work presents a comprehensive benchmark of various methods to assess the groundedness of AI-generated responses, aiming to significantly enhance their reliability. Our experiments include similarity-based metrics and natural language inference models to evaluate whether responses are well-founded in the given contexts. We also explore different prompting strategies for large language models to improve the detection of ungrounded responses. We validated the effectiveness of these methods using a newly created grounding classification corpus, designed specifically for legal queries and corresponding responses from retrieval-augmented prompting, focusing on their alignment with source material. Our results indicate potential in groundedness classification of generated responses, with the best method achieving a macro-F1 score of 0.8. Additionally, we evaluated the methods in terms of their latency to determine their suitability for real-world applications, as this step typically follows the generation process. This capability is essential for processes that may trigger additional manual verification or automated response regeneration. In summary, this study demonstrates the potential of various detection methods to improve the trustworthiness of generative AI in legal settings.
Search
Co-authors
- Dietrich Trautmann 1
- Natalia Ostapuk 1
- Quentin Grail 1
- Guglielmo Bonifazi 1
- Shang Gao 1
- show all...
Venues
- nllp1