LegalLens 2024 Shared Task: Masala-chai Submission

Khalid Rajan, Royal Sequiera


Abstract
In this paper, we present the masala-chai team’s participation in the LegalLens 2024 shared task and detail our approach to predicting legal entities and performing natural language inference (NLI) in the legal domain. We experimented with various transformer-based models, including BERT, RoBERTa, Llama 3.1, and GPT-4o. Our results show that state-of-the-art models like GPT-4o underperformed in NER and NLI tasks, even when using advanced techniques such as bootstrapping and prompt optimization. The best performance in NER (accuracy: 0.806, F1 macro: 0.701) was achieved with a fine-tuned RoBERTa model, while the highest NLI results (accuracy: 0.825, F1 macro: 0.833) came from a fine-tuned Llama 3.1 8B model. Notably, RoBERTa, despite having significantly fewer parameters than Llama 3.1 8B, delivered comparable results. We discuss key findings and insights from our experiments and provide our results and code for reproducibility and further analysis at https://github.com/rosequ/masala-chai
Anthology ID:
2024.nllp-1.30
Volume:
Proceedings of the Natural Legal Language Processing Workshop 2024
Month:
November
Year:
2024
Address:
Miami, FL, USA
Editors:
Nikolaos Aletras, Ilias Chalkidis, Leslie Barrett, Cătălina Goanță, Daniel Preoțiuc-Pietro, Gerasimos Spanakis
Venue:
NLLP
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
346–354
Language:
URL:
https://aclanthology.org/2024.nllp-1.30
DOI:
Bibkey:
Cite (ACL):
Khalid Rajan and Royal Sequiera. 2024. LegalLens 2024 Shared Task: Masala-chai Submission. In Proceedings of the Natural Legal Language Processing Workshop 2024, pages 346–354, Miami, FL, USA. Association for Computational Linguistics.
Cite (Informal):
LegalLens 2024 Shared Task: Masala-chai Submission (Rajan & Sequiera, NLLP 2024)
Copy Citation:
PDF:
https://aclanthology.org/2024.nllp-1.30.pdf