Zero-shot Entity Linking with Less Data

G P Shrivatsa Bhargav, Dinesh Khandelwal, Saswati Dana, Dinesh Garg, Pavan Kapanipathi, Salim Roukos, Alexander Gray, L Venkata Subramaniam


Abstract
Entity Linking (EL) maps an entity mention in a natural language sentence to an entity in a knowledge base (KB). The Zero-shot Entity Linking (ZEL) extends the scope of EL to unseen entities at the test time without requiring new labeled data. BLINK (BERT-based) is one of the SOTA models for ZEL. Interestingly, we discovered that BLINK exhibits diminishing returns, i.e., it reaches 98% of its performance with just 1% of the training data and the remaining 99% of the data yields only a marginal increase of 2% in the performance. While this extra 2% gain makes a huge difference for downstream tasks, training BLINK on large amounts of data is very resource-intensive and impractical. In this paper, we propose a neuro-symbolic, multi-task learning approach to bridge this gap. Our approach boosts the BLINK’s performance with much less data by exploiting an auxiliary information about entity types. Specifically, we train our model on two tasks simultaneously - entity linking (primary task) and hierarchical entity type prediction (auxiliary task). The auxiliary task exploits the hierarchical structure of entity types. Our approach achieves superior performance on ZEL task with significantly less training data. On four different benchmark datasets, we show that our approach achieves significantly higher performance than SOTA models when they are trained with just 0.01%, 0.1%, or 1% of the original training data. Our code is available at https://github.com/IBM/NeSLET.
Anthology ID:
2022.findings-naacl.127
Original:
2022.findings-naacl.127v1
Version 2:
2022.findings-naacl.127v2
Volume:
Findings of the Association for Computational Linguistics: NAACL 2022
Month:
July
Year:
2022
Address:
Seattle, United States
Editors:
Marine Carpuat, Marie-Catherine de Marneffe, Ivan Vladimir Meza Ruiz
Venue:
Findings
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
1681–1697
Language:
URL:
https://aclanthology.org/2022.findings-naacl.127
DOI:
10.18653/v1/2022.findings-naacl.127
Bibkey:
Cite (ACL):
G P Shrivatsa Bhargav, Dinesh Khandelwal, Saswati Dana, Dinesh Garg, Pavan Kapanipathi, Salim Roukos, Alexander Gray, and L Venkata Subramaniam. 2022. Zero-shot Entity Linking with Less Data. In Findings of the Association for Computational Linguistics: NAACL 2022, pages 1681–1697, Seattle, United States. Association for Computational Linguistics.
Cite (Informal):
Zero-shot Entity Linking with Less Data (Bhargav et al., Findings 2022)
Copy Citation:
PDF:
https://aclanthology.org/2022.findings-naacl.127.pdf
Video:
 https://aclanthology.org/2022.findings-naacl.127.mp4
Code
 facebookresearch/BLINK +  additional community code