On the Relationship Between RNN Hidden-State Vectors and Semantic Structures

Edi Muskardin, Martin Tappler, Ingo Pill, Bernhard Aichernig, Thomas Pock


Abstract
We examine the assumption that hidden-state vectors of recurrent neural networks (RNNs) tend to form clusters of semantically similar vectors, which we dub the clustering hypothesis. While this hypothesis has been assumed in RNN analyses in recent years, its validity has not been studied thoroughly on modern RNN architectures. We first consider RNNs that were trained to recognize regular languages. This enables us to draw on perfect ground-truth automata in our evaluation, against which we can compare the RNN’s accuracy and the distribution of the hidden-state vectors. Then, we consider context-free languages to examine if RNN states form clusters for more expressive languages.For our analysis, we fit (generalized) linear models to classify RNN states into automata states and we apply different unsupervised clustering techniques. With a new ambiguity score, derived from information entropy, we measure how well an abstraction function maps the hidden state vectors to abstract clusters. Our evaluation supports the validity of the clustering hypothesis for regular languages, especially if RNNs are well-trained, i.e., clustering techniques succeed in finding clusters of similar state vectors. However, the clustering accuracy decreases substantially for context-free languages. This suggests that clustering is not a reliable abstraction technique for RNNs used in tasks like natural language processing.
Anthology ID:
2024.findings-acl.335
Volume:
Findings of the Association for Computational Linguistics ACL 2024
Month:
August
Year:
2024
Address:
Bangkok, Thailand and virtual meeting
Editors:
Lun-Wei Ku, Andre Martins, Vivek Srikumar
Venue:
Findings
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
5641–5658
Language:
URL:
https://aclanthology.org/2024.findings-acl.335
DOI:
Bibkey:
Cite (ACL):
Edi Muskardin, Martin Tappler, Ingo Pill, Bernhard Aichernig, and Thomas Pock. 2024. On the Relationship Between RNN Hidden-State Vectors and Semantic Structures. In Findings of the Association for Computational Linguistics ACL 2024, pages 5641–5658, Bangkok, Thailand and virtual meeting. Association for Computational Linguistics.
Cite (Informal):
On the Relationship Between RNN Hidden-State Vectors and Semantic Structures (Muskardin et al., Findings 2024)
Copy Citation:
PDF:
https://aclanthology.org/2024.findings-acl.335.pdf