Learning Conceptual Spaces with Disentangled Facets

Rana Alshaikh, Zied Bouraoui, Steven Schockaert


Abstract
Conceptual spaces are geometric representations of meaning that were proposed by G ̈ardenfors (2000). They share many similarities with the vector space embeddings that are commonly used in natural language processing. However, rather than representing entities in a single vector space, conceptual spaces are usually decomposed into several facets, each of which is then modelled as a relatively low dimensional vector space. Unfortunately, the problem of learning such conceptual spaces has thus far only received limited attention. To address this gap, we analyze how, and to what extent, a given vector space embedding can be decomposed into meaningful facets in an unsupervised fashion. While this problem is highly challenging, we show that useful facets can be discovered by relying on word embeddings to group semantically related features.
Anthology ID:
K19-1013
Volume:
Proceedings of the 23rd Conference on Computational Natural Language Learning (CoNLL)
Month:
November
Year:
2019
Address:
Hong Kong, China
Editors:
Mohit Bansal, Aline Villavicencio
Venue:
CoNLL
SIG:
SIGNLL
Publisher:
Association for Computational Linguistics
Note:
Pages:
131–139
Language:
URL:
https://aclanthology.org/K19-1013
DOI:
10.18653/v1/K19-1013
Bibkey:
Cite (ACL):
Rana Alshaikh, Zied Bouraoui, and Steven Schockaert. 2019. Learning Conceptual Spaces with Disentangled Facets. In Proceedings of the 23rd Conference on Computational Natural Language Learning (CoNLL), pages 131–139, Hong Kong, China. Association for Computational Linguistics.
Cite (Informal):
Learning Conceptual Spaces with Disentangled Facets (Alshaikh et al., CoNLL 2019)
Copy Citation:
PDF:
https://aclanthology.org/K19-1013.pdf