Are Data Augmentation Methods in Named Entity Recognition Applicable for Uncertainty Estimation?

Wataru Hashimoto, Hidetaka Kamigaito, Taro Watanabe


Abstract
This work investigates the impact of data augmentation on confidence calibration and uncertainty estimation in Named Entity Recognition (NER) tasks. For the future advance of NER in safety-critical fields like healthcare and finance, it is essential to achieve accurate predictions with calibrated confidence when applying Deep Neural Networks (DNNs), including Pre-trained Language Models (PLMs), as a real-world application. However, DNNs are prone to miscalibration, which limits their applicability. Moreover, existing methods for calibration and uncertainty estimation are computational expensive. Our investigation in NER found that data augmentation improves calibration and uncertainty in cross-genre and cross-lingual setting, especially in-domain setting. Furthermore, we showed that the calibration for NER tends to be more effective when the perplexity of the sentences generated by data augmentation is lower, and that increasing the size of the augmentation further improves calibration and uncertainty.
Anthology ID:
2024.emnlp-main.1049
Volume:
Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing
Month:
November
Year:
2024
Address:
Miami, Florida, USA
Editors:
Yaser Al-Onaizan, Mohit Bansal, Yun-Nung Chen
Venue:
EMNLP
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
18852–18867
Language:
URL:
https://aclanthology.org/2024.emnlp-main.1049
DOI:
Bibkey:
Cite (ACL):
Wataru Hashimoto, Hidetaka Kamigaito, and Taro Watanabe. 2024. Are Data Augmentation Methods in Named Entity Recognition Applicable for Uncertainty Estimation?. In Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing, pages 18852–18867, Miami, Florida, USA. Association for Computational Linguistics.
Cite (Informal):
Are Data Augmentation Methods in Named Entity Recognition Applicable for Uncertainty Estimation? (Hashimoto et al., EMNLP 2024)
Copy Citation:
PDF:
https://aclanthology.org/2024.emnlp-main.1049.pdf