%0 Conference Proceedings %T Quantifying Context Overlap for Training Word Embeddings %A Zhuang, Yimeng %A Xie, Jinghui %A Zheng, Yinhe %A Zhu, Xuan %Y Riloff, Ellen %Y Chiang, David %Y Hockenmaier, Julia %Y Tsujii, Jun’ichi %S Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing %D 2018 %8 oct nov %I Association for Computational Linguistics %C Brussels, Belgium %F zhuang-etal-2018-quantifying %X Most models for learning word embeddings are trained based on the context information of words, more precisely first order co-occurrence relations. In this paper, a metric is designed to estimate second order co-occurrence relations based on context overlap. The estimated values are further used as the augmented data to enhance the learning of word embeddings by joint training with existing neural word embedding models. Experimental results show that better word vectors can be obtained for word similarity tasks and some downstream NLP tasks by the enhanced approach. %R 10.18653/v1/D18-1057 %U https://aclanthology.org/D18-1057 %U https://doi.org/10.18653/v1/D18-1057 %P 587-593