%0 Conference Proceedings %T Rotated Word Vector Representations and their Interpretability %A Park, Sungjoon %A Bak, JinYeong %A Oh, Alice %Y Palmer, Martha %Y Hwa, Rebecca %Y Riedel, Sebastian %S Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing %D 2017 %8 September %I Association for Computational Linguistics %C Copenhagen, Denmark %F park-etal-2017-rotated %X Vector representation of words improves performance in various NLP tasks, but the high dimensional word vectors are very difficult to interpret. We apply several rotation algorithms to the vector representation of words to improve the interpretability. Unlike previous approaches that induce sparsity, the rotated vectors are interpretable while preserving the expressive performance of the original vectors. Furthermore, any prebuilt word vector representation can be rotated for improved interpretability. We apply rotation to skipgrams and glove and compare the expressive power and interpretability with the original vectors and the sparse overcomplete vectors. The results show that the rotated vectors outperform the original and the sparse overcomplete vectors for interpretability and expressiveness tasks. %R 10.18653/v1/D17-1041 %U https://aclanthology.org/D17-1041 %U https://doi.org/10.18653/v1/D17-1041 %P 401-411