Hamada Zahera


2025

pdf bib
LOLA – An Open-Source Massively Multilingual Large Language Model
Nikit Srivastava | Denis Kuchelev | Tatiana Moteu Ngoli | Kshitij Shetty | Michael Roeder | Hamada Zahera | Diego Moussallem | Axel-Cyrille Ngonga Ngomo
Proceedings of the 31st International Conference on Computational Linguistics

This paper presents LOLA, a massively multilingual large language model trained on more than 160 languages using a sparse Mixture-of-Experts Transformer architecture. Our architectural and implementation choices address the challenge of harnessing linguistic diversity while maintaining efficiency and avoiding the common pitfalls of multilinguality. Our analysis of the evaluation results shows competitive performance in natural language generation and understanding tasks. Additionally, we demonstrate how the learned expert-routing mechanism exploits implicit phylogenetic linguistic patterns to potentially alleviate the curse of multilinguality. We provide an in-depth look at the training process, an analysis of the datasets, and a balanced exploration of the model’s strengths and limitations. As an open-source model, LOLA promotes reproducibility and serves as a robust foundation for future research. Our findings enable the development of compute-efficient multilingual models with strong, scalable performance across languages.

pdf bib
Contextual Augmentation for Entity Linking using Large Language Models
Daniel Vollmers | Hamada Zahera | Diego Moussallem | Axel-Cyrille Ngonga Ngomo
Proceedings of the 31st International Conference on Computational Linguistics

Entity Linking involves detecting and linking entity mentions in natural language texts to a knowledge graph. Traditional methods use a two-step process with separate models for entity recognition and disambiguation, which can be computationally intensive and less effective. We propose a fine-tuned model that jointly integrates entity recognition and disambiguation in a unified framework. Furthermore, our approach leverages large language models to enrich the context of entity mentions, yielding better disambiguation. We evaluated our approach on benchmark datasets and compared with several baselines. The evaluation results show that our approach achieves state-of-the-art performance on out-of-domain datasets.