A Survey of Generative Information Extraction

Zikang Zhang, Wangjie You, Tianci Wu, Xinrui Wang, Juntao Li, Min Zhang


Abstract
Generative information extraction (Generative IE) aims to generate structured text sequences from unstructured text using a generative framework. Scaling in model size yields variations in adaptation and generalization, and also drives fundamental shifts in the techniques and approaches used within this domain. In this survey, we first review generative information extraction (IE) methods based on pre-trained language models (PLMs) and large language models (LLMs), focusing on their adaptation and generalization capabilities. We also discuss the connection between these methods and these two aspects. Furthermore, to balance task performance with the substantial computational demands associated with LLMs, we emphasize the importance of model collaboration. Finally, given the advanced capabilities of LLMs, we explore methods for integrating diverse IE tasks into unified models.
Anthology ID:
2025.coling-main.324
Volume:
Proceedings of the 31st International Conference on Computational Linguistics
Month:
January
Year:
2025
Address:
Abu Dhabi, UAE
Editors:
Owen Rambow, Leo Wanner, Marianna Apidianaki, Hend Al-Khalifa, Barbara Di Eugenio, Steven Schockaert
Venue:
COLING
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
4840–4870
Language:
URL:
https://aclanthology.org/2025.coling-main.324/
DOI:
Bibkey:
Cite (ACL):
Zikang Zhang, Wangjie You, Tianci Wu, Xinrui Wang, Juntao Li, and Min Zhang. 2025. A Survey of Generative Information Extraction. In Proceedings of the 31st International Conference on Computational Linguistics, pages 4840–4870, Abu Dhabi, UAE. Association for Computational Linguistics.
Cite (Informal):
A Survey of Generative Information Extraction (Zhang et al., COLING 2025)
Copy Citation:
PDF:
https://aclanthology.org/2025.coling-main.324.pdf