Ivan Sviridov
2026
Multimodal Evaluation of Russian-language Architectures
Artem Chervyakov | Ulyana Isaeva | Anton Emelyanov | Artem Safin | Maria Tikhonova | Alexander Kharitonov | Yulia Lyakh | Petr Surovtsev | Denis Shevelev | Vildan Saburov | Vasily Konovalov | Elisei Rykov | Ivan Sviridov | Amina Miftakhova | Ilseyar Alimova | Alexander Panchenko | Alexander Kapitanov | Alena Fenogenova
Proceedings of the 19th Conference of the European Chapter of the Association for Computational Linguistics (Volume 1: Long Papers)
Artem Chervyakov | Ulyana Isaeva | Anton Emelyanov | Artem Safin | Maria Tikhonova | Alexander Kharitonov | Yulia Lyakh | Petr Surovtsev | Denis Shevelev | Vildan Saburov | Vasily Konovalov | Elisei Rykov | Ivan Sviridov | Amina Miftakhova | Ilseyar Alimova | Alexander Panchenko | Alexander Kapitanov | Alena Fenogenova
Proceedings of the 19th Conference of the European Chapter of the Association for Computational Linguistics (Volume 1: Long Papers)
Multimodal large language models (MLLMs) are currently at the center of research attention, showing rapid progress in scale and capabilities, yet their intelligence, limitations, and risks remain insufficiently understood. To address these issues, particularly in the context of the Russian language, where no multimodal benchmarks currently exist, we introduce MERA Multi, an open multimodal evaluation framework for Russian-spoken architectures. The benchmark is instruction-based and encompasses default text, image, audio, and video modalities, comprising 18 newly constructed evaluation tasks for both general-purpose models and modality-specific architectures (image-to-text, video-to-text, and audio-to-text). Our contributions include: (i) a universal taxonomy of multimodal abilities; (ii) 18 datasets created entirely from scratch with attention to Russian cultural and linguistic specificity, unified prompts, and metrics; (iii) baseline results for both closed-source and open-source models; (iv) a methodology for preventing benchmark leakage, including watermarking for private sets. While our current focus is on Russian, the proposed benchmark provides a replicable methodology for constructing multimodal benchmarks in typologically diverse languages, particularly within the Slavic language family.
2025
RuCCoD: Towards Automated ICD Coding in Russian
Alexandr Nesterov | Andrey Sakhovskiy | Ivan Sviridov | Airat Valiev | Vladimir Makharev | Petr Anokhin | Galina Zubkova | Elena Tutubalina
Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing
Alexandr Nesterov | Andrey Sakhovskiy | Ivan Sviridov | Airat Valiev | Vladimir Makharev | Petr Anokhin | Galina Zubkova | Elena Tutubalina
Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing
This study investigates the feasibility of automating clinical coding in Russian, a language with limited biomedical resources. We present a new dataset for ICD coding, which includes diagnosis fields from electronic health records (EHRs) annotated with over 10,000 entities and more than 1,500 unique ICD codes. This dataset serves as a benchmark for several state-of-the-art models, including BERT, LLaMA with LoRA, and RAG, with additional experiments examining transfer learning across domains (from PubMed abstracts to medical diagnosis) and terminologies (from UMLS concepts to ICD codes). We then apply the best-performing model to label an in-house EHR dataset containing patient histories from 2017 to 2021. Our experiments, conducted on a carefully curated test set, demonstrate that training with the automated predicted codes leads to a significant improvement in accuracy compared to manually annotated data from physicians. We believe our findings offer valuable insights into the potential for automating clinical coding in resource-limited languages like Russian, which could enhance clinical efficiency and data accuracy in these contexts. Our code and dataset are available at https://github.com/auto-icd-coding/ruccod.
3MDBench: Medical Multimodal Multi-agent Dialogue Benchmark
Ivan Sviridov | Amina Miftakhova | Artemiy Tereshchenko | Galina Zubkova | Pavel Blinov | Andrey Savchenko
Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing
Ivan Sviridov | Amina Miftakhova | Artemiy Tereshchenko | Galina Zubkova | Pavel Blinov | Andrey Savchenko
Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing
Though Large Vision-Language Models (LVLMs) are being actively explored in medicine, their ability to conduct complex real-world telemedicine consultations combining accurate diagnosis with professional dialogue remains underexplored. This paper presents 3MDBench (Medical Multimodal Multi-agent Dialogue Benchmark), an open-source framework for simulating and evaluating LVLM-driven telemedical consultations. 3MDBench simulates patient variability through temperament-based Patient Agent and evaluates diagnostic accuracy and dialogue quality via Assessor Agent. It includes 2996 cases across 34 diagnoses from real-world telemedicine interactions, combining textual and image-based data. The experimental study compares diagnostic strategies for widely used open and closed-source LVLMs. We demonstrate that multimodal dialogue with internal reasoning improves F1 score by 6.5% over non-dialogue settings, highlighting the importance of context-aware, information-seeking questioning. Moreover, injecting predictions from a diagnostic convolutional neural network into the LVLM’s context boosts F1 by up to 20%. Source code is available at https://github.com/univanxx/3mdbench.
Search
Fix author
Co-authors
- Amina Miftakhova 2
- Galina Zubkova 2
- Ilseyar Alimova 1
- Petr Anokhin 1
- Pavel Blinov 1
- Artem Chervyakov 1
- Anton Emelyanov 1
- Alena Fenogenova 1
- Ulyana Isaeva 1
- Alexander Kapitanov 1
- Alexander Kharitonov 1
- Vasily Konovalov 1
- Yulia Lyakh 1
- Vladimir Makharev 1
- Alexandr Nesterov 1
- Alexander Panchenko 1
- Elisei Rykov 1
- Vildan Saburov 1
- Artem Safin 1
- Andrey Sakhovskiy 1
- Andrey Savchenko 1
- Denis Shevelev 1
- Petr Surovtsev 1
- Artemiy Tereshchenko 1
- Maria Tikhonova 1
- Elena Tutubalina 1
- Airat Valiev 1