SmartSpanNER: Making SpanNER Robust in Low Resource Scenarios

Min Zhang, Xiaosong Qiao, Yanqing Zhao, Shimin Tao, Hao Yang


Abstract
Named Entity Recognition (NER) is one of the most fundamental tasks in natural language processing. Span-level prediction (SpanNER) is more naturally suitable for nested NER than sequence labeling (SeqLab). However, according to our experiments, the SpanNER method is more sensitive to the amount of training data, i.e., the F1 score of SpanNER drops much more than that of SeqLab when the amount of training data drops. In order to improve the robustness of SpanNER in low resource scenarios, we propose a simple and effective method SmartSpanNER, which introduces a Named Entity Head (NEH) prediction task to SpanNER and performs multi-task learning together with the task of span classification. Experimental results demonstrate that the robustness of SpanNER could be greatly improved by SmartSpanNER in low resource scenarios constructed on the CoNLL03, Few-NERD, GENIA and ACE05 standard benchmark datasets.
Anthology ID:
2023.findings-emnlp.535
Volume:
Findings of the Association for Computational Linguistics: EMNLP 2023
Month:
December
Year:
2023
Address:
Singapore
Editors:
Houda Bouamor, Juan Pino, Kalika Bali
Venue:
Findings
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
7964–7976
Language:
URL:
https://aclanthology.org/2023.findings-emnlp.535
DOI:
10.18653/v1/2023.findings-emnlp.535
Bibkey:
Cite (ACL):
Min Zhang, Xiaosong Qiao, Yanqing Zhao, Shimin Tao, and Hao Yang. 2023. SmartSpanNER: Making SpanNER Robust in Low Resource Scenarios. In Findings of the Association for Computational Linguistics: EMNLP 2023, pages 7964–7976, Singapore. Association for Computational Linguistics.
Cite (Informal):
SmartSpanNER: Making SpanNER Robust in Low Resource Scenarios (Zhang et al., Findings 2023)
Copy Citation:
PDF:
https://aclanthology.org/2023.findings-emnlp.535.pdf