Word Embeddings as Features for Supervised Coreference Resolution

Iliana Simova, Hans Uszkoreit


Abstract
A common reason for errors in coreference resolution is the lack of semantic information to help determine the compatibility between mentions referring to the same entity. Distributed representations, which have been shown successful in encoding relatedness between words, could potentially be a good source of such knowledge. Moreover, being obtained in an unsupervised manner, they could help address data sparsity issues in labeled training data at a small cost. In this work we investigate whether and to what extend features derived from word embeddings can be successfully used for supervised coreference resolution. We experiment with several word embedding models, and several different types of embeddingbased features, including embedding cluster and cosine similarity-based features. Our evaluations show improvements in the performance of a supervised state-of-theart coreference system.
Anthology ID:
R17-1088
Volume:
Proceedings of the International Conference Recent Advances in Natural Language Processing, RANLP 2017
Month:
September
Year:
2017
Address:
Varna, Bulgaria
Editors:
Ruslan Mitkov, Galia Angelova
Venue:
RANLP
SIG:
Publisher:
INCOMA Ltd.
Note:
Pages:
686–693
Language:
URL:
https://doi.org/10.26615/978-954-452-049-6_088
DOI:
10.26615/978-954-452-049-6_088
Bibkey:
Cite (ACL):
Iliana Simova and Hans Uszkoreit. 2017. Word Embeddings as Features for Supervised Coreference Resolution. In Proceedings of the International Conference Recent Advances in Natural Language Processing, RANLP 2017, pages 686–693, Varna, Bulgaria. INCOMA Ltd..
Cite (Informal):
Word Embeddings as Features for Supervised Coreference Resolution (Simova & Uszkoreit, RANLP 2017)
Copy Citation:
PDF:
https://doi.org/10.26615/978-954-452-049-6_088