Leon Engländer
2024
M2QA: Multi-domain Multilingual Question Answering
Leon Engländer
|
Hannah Sterz
|
Clifton A Poth
|
Jonas Pfeiffer
|
Ilia Kuznetsov
|
Iryna Gurevych
Findings of the Association for Computational Linguistics: EMNLP 2024
Generalization and robustness to input variation are core desiderata of machine learning research. Language varies along several axes, most importantly, language instance (e.g. French) and domain (e.g. news). While adapting NLP models to new languages within a single domain, or to new domains within a single language, is widely studied, research in joint adaptation is hampered by the lack of evaluation datasets. This prevents the transfer of NLP systems from well-resourced languages and domains to non-dominant language-domain combinations. To address this gap, we introduce M2QA, a multi-domain multilingual question answering benchmark.M2QA includes 13,500 SQuAD 2.0-style question-answer instances in German, Turkish, and Chinese for the domains of product reviews, news, and creative writing. We use M2QA to explore cross-lingual cross-domain performance of fine-tuned models and state-of-the-art LLMs and investigate modular approaches to domain and language adaptation.We witness **1)** considerable performance _variations_ across domain-language combinations within model classes and **2)** considerable performance _drops_ between source and target language-domain combinations across all model sizes. We demonstrate that M2QA is far from solved, and new methods to effectively transfer both linguistic and domain-specific information are necessary.
2023
Adapters: A Unified Library for Parameter-Efficient and Modular Transfer Learning
Clifton Poth
|
Hannah Sterz
|
Indraneil Paul
|
Sukannya Purkayastha
|
Leon Engländer
|
Timo Imhof
|
Ivan Vulić
|
Sebastian Ruder
|
Iryna Gurevych
|
Jonas Pfeiffer
Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing: System Demonstrations
We introduce Adapters, an open-source library that unifies parameter-efficient and modular transfer learning in large language models. By integrating 10 diverse adapter methods into a unified interface, Adapters offers ease of use and flexible configuration. Our library allows researchers and practitioners to leverage adapter modularity through composition blocks, enabling the design of complex adapter setups. We demonstrate the library’s efficacy by evaluating its performance against full fine-tuning on various NLP tasks. Adapters provides a powerful tool for addressing the challenges of conventional fine-tuning paradigms and promoting more efficient and modular transfer learning. The library is available via https://adapterhub.ml/adapters.
Search
Co-authors
- Hannah Sterz 2
- Iryna Gurevych 2
- Jonas Pfeiffer 2
- Clifton Poth 1
- Indraneil Paul 1
- show all...