Sergei Tilga
2026
JEEM: Vision-Language Understanding in Four Arabic Dialects
Karima Kadaoui | Hanin Atwany | Hamdan Al-Ali | Abdelrahman Mohamed | Ali Mekky | Sergei Tilga | Natalia Fedorova | Ekaterina Artemova | Hanan Aldarmaki | Yova Kementchedjhieva
Findings of the Association for Computational Linguistics: EACL 2026
Karima Kadaoui | Hanin Atwany | Hamdan Al-Ali | Abdelrahman Mohamed | Ali Mekky | Sergei Tilga | Natalia Fedorova | Ekaterina Artemova | Hanan Aldarmaki | Yova Kementchedjhieva
Findings of the Association for Computational Linguistics: EACL 2026
We introduce JEEM, a benchmark designed to evaluate Vision-Language Models (VLMs) on visual understanding across four Arabic-speaking countries: Jordan, The Emirates, Egypt, and Morocco. JEEM includes the tasks of image captioning and visual question answering, and features culturally rich and regionally diverse content. This dataset aims to assess the ability of VLMs to generalize across dialects and accurately interpret cultural elements in visual contexts. In an evaluation of five prominent open-source Arabic VLMs and GPT-4o, we find that the Arabic VLMs consistently underperform, struggling with both visual understanding and dialect-specific generation. While GPT-4o ranks best in this comparison, the model’s linguistic competence varies across dialects, and its visual understanding capabilities lag behind. This underscores the need for more inclusive models and the value of culturally-diverse evaluation paradigms.
2025
U-MATH: A University-Level Benchmark for Evaluating Mathematical Skills in Large Language Models
Konstantin Chernyshev | Vitaliy Polshkov | Vlad Stepanov | Alex Myasnikov | Ekaterina Artemova | Alexei Miasnikov | Sergei Tilga
Proceedings of the Fourth Workshop on Generation, Evaluation and Metrics (GEM²)
Konstantin Chernyshev | Vitaliy Polshkov | Vlad Stepanov | Alex Myasnikov | Ekaterina Artemova | Alexei Miasnikov | Sergei Tilga
Proceedings of the Fourth Workshop on Generation, Evaluation and Metrics (GEM²)
Current evaluations of mathematical skills in Large Language Models are constrained by benchmarks lacking scope, particularly for multi-modal problems — frequently relying on school-level, niche Olympiad-style, simple quiz-format, or relatively small datasets.To address this, we introduce **U-MATH**, a novel benchmark comprising **1,100** unpublished open-ended university-level problems sourced from current US curricula, with **20%** incorporating visual elements. Given the free-form nature of U-MATH problems, we employ LLM judges for solution evaluation and release 𝜇**-MATH**, a meta-evaluation benchmark composed of **1,084** U-MATH-derived tasks enabling precise assessment of these judges.Benchmarking leading LLMs reveals marked limitations in multi-modal reasoning, with maximum accuracy reaching 93.1% on textual tasks but only 58.5% on visual ones. Furthermore, solution judgment proves challenging, requiring the most advanced models to achieve meaningfully high performance, even still peaking at an imperfect F1-score of 90.1%.
Beemo: Benchmark of Expert-edited Machine-generated Outputs
Ekaterina Artemova | Jason S Lucas | Saranya Venkatraman | Jooyoung Lee | Sergei Tilga | Adaku Uchendu | Vladislav Mikhailov
Proceedings of the 2025 Conference of the Nations of the Americas Chapter of the Association for Computational Linguistics: Human Language Technologies (Volume 1: Long Papers)
Ekaterina Artemova | Jason S Lucas | Saranya Venkatraman | Jooyoung Lee | Sergei Tilga | Adaku Uchendu | Vladislav Mikhailov
Proceedings of the 2025 Conference of the Nations of the Americas Chapter of the Association for Computational Linguistics: Human Language Technologies (Volume 1: Long Papers)
The rapid proliferation of large language models (LLMs) has increased the volume of machine-generated texts (MGTs) and blurred text authorship in various domains. However, most existing MGT benchmarks include single-author texts (human-written and machine-generated). This conventional design fails to capture more practical multi-author scenarios, where the user refines the LLM response for natural flow, coherence, and factual correctness. Our paper introduces the Benchmark of Expert-edited Machine-generated Outputs (Beemo), which includes 6.5k texts written by humans, generated by ten instruction-finetuned LLMs, and edited by experts for various use cases, ranging from creative writing to summarization. Beemo additionally comprises 13.1k machine-generated and LLM-edited texts, allowing for diverse MGT detection evaluation across various edit types. We document Beemo’s creation protocol and present the results of benchmarking 33 configurations of MGT detectors in different experimental setups. We find that expert-based editing evades MGT detection, while LLM-edited texts are unlikely to be recognized as human-written. Beemo and all materials are publicly available.