Language Modeling Using Entanglement Enhanced Tensor Trains

Ellis Reyes, Yi-Shin Chen


Abstract
Tensor Train Language Models (TTLMs) offer significant memory savings by representing text sequences as tensor networks, but naive implementations struggle with long-range dependencies and limited flexibility. We introduce a modular TTLM framework that combine local and non-local context modules to achieve scalable language modeling. Our non-local modules, inspired by entanglement in quantum information theory, enable efficient modeling of long-range interactions between hidden states. Experiments on Penn Treebank and Wikitext datasets show that our modular TTLM, including entanglement-augmented variants, outperform naive baselines. These results highlight TTLMs as a promising, memory-efficient alternatives for modern language modeling.
Anthology ID:
2025.rocling-main.27
Volume:
Proceedings of the 37th Conference on Computational Linguistics and Speech Processing (ROCLING 2025)
Month:
November
Year:
2025
Address:
National Taiwan University, Taipei City, Taiwan
Editors:
Kai-Wei Chang, Ke-Han Lu, Chih-Kai Yang, Zhi-Rui Tam, Wen-Yu Chang, Chung-Che Wang
Venue:
ROCLING
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
258–265
Language:
URL:
https://aclanthology.org/2025.rocling-main.27/
DOI:
Bibkey:
Cite (ACL):
Ellis Reyes and Yi-Shin Chen. 2025. Language Modeling Using Entanglement Enhanced Tensor Trains. In Proceedings of the 37th Conference on Computational Linguistics and Speech Processing (ROCLING 2025), pages 258–265, National Taiwan University, Taipei City, Taiwan. Association for Computational Linguistics.
Cite (Informal):
Language Modeling Using Entanglement Enhanced Tensor Trains (Reyes & Chen, ROCLING 2025)
Copy Citation:
PDF:
https://aclanthology.org/2025.rocling-main.27.pdf