Tombekai Vangoni Sherman
2025
IrokoBench: A New Benchmark for African Languages in the Age of Large Language Models
David Ifeoluwa Adelani | Jessica Ojo | Israel Abebe Azime | Jian Yun Zhuang | Jesujoba Oluwadara Alabi | Xuanli He | Millicent Ochieng | Sara Hooker | Andiswa Bukula | En-Shiun Annie Lee | Chiamaka Ijeoma Chukwuneke | Happy Buzaaba | Blessing Kudzaishe Sibanda | Godson Koffi Kalipe | Jonathan Mukiibi | Salomon Kabongo Kabenamualu | Foutse Yuehgoh | Mmasibidi Setaka | Lolwethu Ndolela | Nkiruka Odu | Rooweither Mabuya | Salomey Osei | Shamsuddeen Hassan Muhammad | Sokhar Samb | Tadesse Kebede Guge | Tombekai Vangoni Sherman | Pontus Stenetorp
Proceedings of the 2025 Conference of the Nations of the Americas Chapter of the Association for Computational Linguistics: Human Language Technologies (Volume 1: Long Papers)
David Ifeoluwa Adelani | Jessica Ojo | Israel Abebe Azime | Jian Yun Zhuang | Jesujoba Oluwadara Alabi | Xuanli He | Millicent Ochieng | Sara Hooker | Andiswa Bukula | En-Shiun Annie Lee | Chiamaka Ijeoma Chukwuneke | Happy Buzaaba | Blessing Kudzaishe Sibanda | Godson Koffi Kalipe | Jonathan Mukiibi | Salomon Kabongo Kabenamualu | Foutse Yuehgoh | Mmasibidi Setaka | Lolwethu Ndolela | Nkiruka Odu | Rooweither Mabuya | Salomey Osei | Shamsuddeen Hassan Muhammad | Sokhar Samb | Tadesse Kebede Guge | Tombekai Vangoni Sherman | Pontus Stenetorp
Proceedings of the 2025 Conference of the Nations of the Americas Chapter of the Association for Computational Linguistics: Human Language Technologies (Volume 1: Long Papers)
Despite the widespread adoption of Large language models (LLMs), their remarkable capabilities remain limited to a few high-resource languages. Additionally, many low-resource languages (e.g. African languages) are often evaluated only on basic text classification tasks due to the lack of appropriate or comprehensive benchmarks outside of high-resource languages. In this paper, we introduce IrokoBench—a human-translated benchmark dataset for 17 typologically-diverse low-resource African languages covering three tasks: natural language inference(AfriXNLI), mathematical reasoning(AfriMGSM), and multi-choice knowledge-based QA(AfriMMLU). We use IrokoBench to evaluate zero-shot, few-shot, and translate-test settings(where test sets are translated into English) across 10 open and four proprietary LLMs. Our evaluation reveals a significant performance gap between high-resource languages (such as English and French) and low-resource African languages. We observe a significant performance gap between open and proprietary models, with the highest performing open model, Gemma 2 27B only at 63% of the best-performing proprietary model GPT-4o performance. Machine translating the test set to English before evaluation helped to close the gap for larger models that are English-centric, like Gemma 2 27B and LLaMa 3.1 70B. These findings suggest that more efforts are needed to develop and adapt LLMs for African languages.
Search
Fix author
Co-authors
- David Ifeoluwa Adelani 1
- Jesujoba Alabi 1
- Israel Abebe Azime 1
- Andiswa Bukula 1
- Happy Buzaaba 1
- Chiamaka Ijeoma Chukwuneke 1
- Tadesse Kebede Guge 1
- Xuanli He 1
- Sara Hooker 1
- Salomon Kabongo Kabenamualu 1
- Godson Koffi Kalipe 1
- En-Shiun Annie Lee 1
- Rooweither Mabuya 1
- Shamsuddeen Hassan Muhammad 1
- Jonathan Mukiibi 1
- Lolwethu Ndolela 1
- Millicent Ochieng 1
- Nkiruka Odu 1
- Jessica Ojo 1
- Salomey Osei 1
- Sokhar Samb 1
- Mmasibidi Setaka 1
- Blessing Kudzaishe Sibanda 1
- Pontus Stenetorp 1
- Foutse Yuehgoh 1
- Jian Yun Zhuang 1