Wantong Zhao
2025
Beyond Boundaries: Learning a Universal Entity Taxonomy across Datasets and Languages for Open Named Entity Recognition
Yuming Yang
|
Wantong Zhao
|
Caishuang Huang
|
Junjie Ye
|
Xiao Wang
|
Huiyuan Zheng
|
Yang Nan
|
Yuran Wang
|
Xueying Xu
|
Kaixin Huang
|
Yunke Zhang
|
Tao Gui
|
Qi Zhang
|
Xuanjing Huang
Proceedings of the 31st International Conference on Computational Linguistics
Open Named Entity Recognition (NER), which involves identifying arbitrary types of entities from arbitrary domains, remains challenging for Large Language Models (LLMs). Recent studies suggest that fine-tuning LLMs on extensive NER data can boost their performance. However, training directly on existing datasets neglects their inconsistent entity definitions and redundant data, limiting LLMs to dataset-specific learning and hindering out-of-domain adaptation. To address this, we present B2NERD, a compact dataset designed to guide LLMs’ generalization in Open NER under a universal entity taxonomy. B2NERD is refined from 54 existing English and Chinese datasets using a two-step process. First, we detect inconsistent entity definitions across datasets and clarify them by distinguishable label names to construct a universal taxonomy of 400+ entity types. Second, we address redundancy using a data pruning strategy that selects fewer samples with greater category and semantic diversity. Comprehensive evaluation shows that B2NERD significantly enhances LLMs’ Open NER capabilities. Our B2NER models, trained on B2NERD, outperform GPT-4 by 6.8-12.0 F1 points and surpass previous methods in 3 out-of-domain benchmarks across 15 datasets and 6 languages. The data, models, and code are publicly available at https://github.com/UmeanNever/B2NER.
Search
Fix data
Co-authors
- Tao Gui 1
- Caishuang Huang 1
- Kaixin Huang 1
- Xuan-Jing Huang (黄萱菁) 1
- Yang Nan 1
- show all...