Kelechi Ezema
2026
AfriVox: Probing Multilingual and Accent Robustness of Speech LLMs
Busayo Awobade | Mardhiyah Sanni | Tassallah Abdullahi | Chibuzor Okocha | Kelechi Ezema | Devendra Deepak Kayande | Lukman Enegi Ismaila | Tobi Olatunji | Gloria Ashiya Katuka
Proceedings of the 19th Conference of the European Chapter of the Association for Computational Linguistics (Volume 1: Long Papers)
Busayo Awobade | Mardhiyah Sanni | Tassallah Abdullahi | Chibuzor Okocha | Kelechi Ezema | Devendra Deepak Kayande | Lukman Enegi Ismaila | Tobi Olatunji | Gloria Ashiya Katuka
Proceedings of the 19th Conference of the European Chapter of the Association for Computational Linguistics (Volume 1: Long Papers)
Recent advances in multimodal and speech-native large language models (LLMs) have delivered impressive speech recognition, translation, understanding, and question-answering capabilities for high-resource languages. However, African languages and non-native French or English accents remain dramatically underrepresented in benchmarks limiting the understanding and applicability of leading LLMs for millions of francophone and anglophone users in low-resource settings. We presents AfriVox, an open-source benchmark (including novel domain-specific and unscripted datasets) across 20 African languages, African-accented French, Arabic, and 100+ African English accents, contrasting leading multimodal speech LLMs with traditional unimodal automatic speech transcription (ASR) and translation (AST) models. Our analysis reveals significant language coverage variation, surprising LLM translation performance gains (e.g. Gemini), robustness concerns with unscripted speech, and substantial performance disparities for "supported" African languages. We profile the strengths, limitations, and language support of each model, and conduct the first targeted fine-tuning of a modern speech LLM (Qwen2.5-Omni) for three Nigerian languages, exceeding SOTA, and achieving up to 54% relative WER reduction and significant BLEU gains, offering practical guidance for implementers seeking to serve local language users.
AfriMMT-EA: Multi-domain Machine Translation for Low-Resource East African Languages
Naome A Etori | Kelechi Ezema | Nathaniel Romney Robinson | Davis David | Alfred Malengo Kondoro | Elisha Ondieki Makori | Michael Samwel Mollel | Maria Gini
Findings of the Association for Computational Linguistics: EACL 2026
Naome A Etori | Kelechi Ezema | Nathaniel Romney Robinson | Davis David | Alfred Malengo Kondoro | Elisha Ondieki Makori | Michael Samwel Mollel | Maria Gini
Findings of the Association for Computational Linguistics: EACL 2026
Despite remarkable progress in multilingual machine translation (MT), the majority of African—especially East African—languages remain significantly underrepresented both in benchmark datasets and state-of-the-art (SOTA) MT models. This persistent exclusion from mainstream technologies not only limits equitable access, but constrains the development of tools that accurately reflect the region’s linguistic and cultural diversity. Recent advances in open-source large language models have demonstrated strong multilingual MT capabilities through data-efficient adaptation strategies. However, little work has explored their potential for low-resource African languages. We introduce AfriMMT-EA, the first highly multilingual benchmark and MT dataset for East African languages. Our datasets comprise 54 local languages across five East African countries. We used these data to fine-tune two multilingual versions of Gemma-3. We compare models’ performance on these languages with larger off-the-shelf baselines. We release our data and models, in the interest of advancing MT for these low-resource languages and their communities.
2025
Linguistic Alignment Predicts Learning in Small Group Tutoring Sessions
Dorothea French | Robert Moulder | Kelechi Ezema | Katharina von der Wense | Sidney K. DMello
Findings of the Association for Computational Linguistics: EMNLP 2025
Dorothea French | Robert Moulder | Kelechi Ezema | Katharina von der Wense | Sidney K. DMello
Findings of the Association for Computational Linguistics: EMNLP 2025
Cognitive science offers rich theories of learning and communication, yet these are often difficult to operationalize at scale. We demonstrate how natural language processing can bridge this gap by applying psycholinguistic theories of discourse to real-world educational data. We investigate linguistic alignment – the convergence of conversational partners’ word choice, grammar, and meaning – in a longitudinal dataset of real-world tutoring interactions and associated student test scores. We examine (1) the extent of alignment, (2) role-based patterns among tutors and students, and (3) the relationship between alignment and learning outcomes. We find that both tutors and students exhibit lexical, syntactic, and semantic alignment, with tutors aligning more strongly to students. Crucially, tutor lexical alignment predicts student learning gains, while student lexical alignment negatively predicts them. As a lightweight, interpretable metric, linguistic alignment offers practical applications in intelligent tutoring systems, educator dashboards, and tutor training.