Evaluating Open-Source LLMs in Low-Resource Languages: Insights from Latvian High School Exams

Roberts Darģis, Guntis Bārzdiņš, Inguna Skadiņa, Baiba Saulite


Abstract
The latest large language models (LLM) have significantly advanced natural language processing (NLP) capabilities across various tasks. However, their performance in low-resource languages, such as Latvian with 1.5 million native speakers, remains substantially underexplored due to both limited training data and the absence of comprehensive evaluation benchmarks. This study addresses this gap by conducting a systematic assessment of prominent open-source LLMs on natural language understanding (NLU) and natural language generation (NLG) tasks in Latvian. We utilize standardized high school centralized graduation exams as a benchmark dataset, offering relatable and diverse evaluation scenarios that encompass multiple-choice questions and complex text analysis tasks. Our experimental setup involves testing models from the leading LLM families, including Llama, Qwen, Gemma, and Mistral, with OpenAI’s GPT-4 serving as a performance reference. The results reveal that certain open-source models demonstrate competitive performance in NLU tasks, narrowing the gap with GPT-4. However, all models exhibit notable deficiencies in NLG tasks, specifically in generating coherent and contextually appropriate text analyses, highlighting persistent challenges in NLG for low-resource languages. These findings contribute to efforts to develop robust multilingual benchmarks and improve LLM performance in diverse linguistic contexts.
Anthology ID:
2024.nlp4dh-1.28
Volume:
Proceedings of the 4th International Conference on Natural Language Processing for Digital Humanities
Month:
November
Year:
2024
Address:
Miami, USA
Editors:
Mika Hämäläinen, Emily Öhman, So Miyagawa, Khalid Alnajjar, Yuri Bizzoni
Venue:
NLP4DH
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
289–293
Language:
URL:
https://aclanthology.org/2024.nlp4dh-1.28
DOI:
Bibkey:
Cite (ACL):
Roberts Darģis, Guntis Bārzdiņš, Inguna Skadiņa, and Baiba Saulite. 2024. Evaluating Open-Source LLMs in Low-Resource Languages: Insights from Latvian High School Exams. In Proceedings of the 4th International Conference on Natural Language Processing for Digital Humanities, pages 289–293, Miami, USA. Association for Computational Linguistics.
Cite (Informal):
Evaluating Open-Source LLMs in Low-Resource Languages: Insights from Latvian High School Exams (Darģis et al., NLP4DH 2024)
Copy Citation:
PDF:
https://aclanthology.org/2024.nlp4dh-1.28.pdf