Improving Dialect Robustness in Large Language Models via LoRA and Mixture-of-Experts

Sanjh Maheshwari, Aniket Singh Rajpoot, Oana Cocarascu, Mamta .


Abstract
Despite the success of large language models (LLMs) in a wide range of applications, it has been shown that their performance varies across English dialects. Differences among English dialects are reflected in vocabulary, syntax, and writing style, and can adversely affect model performance. Several studies evaluate the dialect robustness of LLMs, yet research on enhancing their robustness to dialectal variation remains limited. In this paper, we propose two parameter-efficient frameworks for improving dialectal robustness in LLMs: DialectFusion where we train separate LoRA layers for each dialect and apply different LoRA merging methods, and DialectMoE which is built on top of Mixture of Experts LoRA and introduces multiple LoRA-based experts to the feed-forward layer to internally model the dialectal dependencies. Our comprehensive analysis on five open-source LLMs for sentiment and sarcasm tasks in zero- and few-shot settings shows that our proposed approaches enhance the dialect robustness of LLMs and outperforms instruct and LoRA fine-tuning based approaches.
Anthology ID:
2026.vardial-1.24
Volume:
Proceedings of the 13th Workshop on NLP for Similar Languages, Varieties and Dialects
Month:
March
Year:
2026
Address:
Rabat, Morocco
Venues:
VarDial | WS
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
293–303
Language:
URL:
https://aclanthology.org/2026.vardial-1.24/
DOI:
Bibkey:
Cite (ACL):
Sanjh Maheshwari, Aniket Singh Rajpoot, Oana Cocarascu, and Mamta .. 2026. Improving Dialect Robustness in Large Language Models via LoRA and Mixture-of-Experts. In Proceedings of the 13th Workshop on NLP for Similar Languages, Varieties and Dialects, pages 293–303, Rabat, Morocco. Association for Computational Linguistics.
Cite (Informal):
Improving Dialect Robustness in Large Language Models via LoRA and Mixture-of-Experts (Maheshwari et al., VarDial 2026)
Copy Citation:
PDF:
https://aclanthology.org/2026.vardial-1.24.pdf