Defne Circi
2025
Evaluating Morphological Compositional Generalization in Large Language Models
Mete Ismayilzada
|
Defne Circi
|
Jonne Sälevä
|
Hale Sirin
|
Abdullatif Köksal
|
Bhuwan Dhingra
|
Antoine Bosselut
|
Duygu Ataman
|
Lonneke Van Der Plas
Proceedings of the 2025 Conference of the Nations of the Americas Chapter of the Association for Computational Linguistics: Human Language Technologies (Volume 1: Long Papers)
Large language models (LLMs) have demonstrated significant progress in various natural language generation and understanding tasks. However, their linguistic generalization capabilities remain questionable, raising doubts about whether these models learn language similarly to humans. While humans exhibit compositional generalization and linguistic creativity in language use, the extent to which LLMs replicate these abilities, particularly in morphology, is under-explored. In this work, we systematically investigate the morphological generalization abilities of LLMs through the lens of compositionality. We define morphemes as compositional primitives and design a novel suite of generative and discriminative tasks to assess morphological productivity and systematicity. Focusing on agglutinative languages such as Turkish and Finnish, we evaluate several state-of-the-art instruction-finetuned multilingual models, including GPT-4 and Gemini. Our analysis shows that LLMs struggle with morphological compositional generalization particularly when applied to novel word roots, with performance declining sharply as morphological complexity increases. While models can identify individual morphological combinations better than chance, their performance lacks systematicity, leading to significant accuracy gaps compared to humans.
2024
Extracting Polymer Nanocomposite Samples from Full-Length Documents
Ghazal Khalighinejad
|
Defne Circi
|
L. Brinson
|
Bhuwan Dhingra
Findings of the Association for Computational Linguistics: ACL 2024
This paper investigates the use of large language models (LLMs) for extracting sample lists of polymer nanocomposites (PNCs) from full-length materials science research papers. The challenge lies in the complex nature of PNC samples, which have numerous attributes scattered throughout the text. The complexity of annotating detailed information on PNCs limits the availability of data, making conventional document-level relation extraction techniques impractical due to the challenge in creating comprehensive named entity span annotations.To address this, we introduce a new benchmark and an evaluation technique for this task and explore different prompting strategies in a zero-shot manner. We also incorporate self-consistency to improve the performance. Our findings show that even advanced LLMs struggle to extract all of the samples from an article. Finally, we analyze the errors encountered in this process, categorizing them into three main challenges, and discuss potential strategies for future research to overcome them.
Search
Fix data
Co-authors
- Bhuwan Dhingra 2
- Duygu Ataman 1
- Antoine Bosselut 1
- L. Brinson 1
- Mete Ismayilzada 1
- show all...