%0 Conference Proceedings %T More Than Words: Collocation Retokenization for Latent Dirichlet Allocation Models %A Cheevaprawatdomrong, Jin %A Schofield, Alexandra %A Rutherford, Attapol %Y Muresan, Smaranda %Y Nakov, Preslav %Y Villavicencio, Aline %S Findings of the Association for Computational Linguistics: ACL 2022 %D 2022 %8 May %I Association for Computational Linguistics %C Dublin, Ireland %F cheevaprawatdomrong-etal-2022-words %X Traditionally, Latent Dirichlet Allocation (LDA) ingests words in a collection of documents to discover their latent topics using word-document co-occurrences. Previous studies show that representing bigrams collocations in the input can improve topic coherence in English. However, it is unclear how to achieve the best results for languages without marked word boundaries such as Chinese and Thai. Here, we explore the use of retokenization based on chi-squared measures, t-statistics, and raw frequency to merge frequent token ngrams into collocations when preparing input to the LDA model. Based on the goodness of fit and the coherence metric, we show that topics trained with merged tokens result in topic keys that are clearer, more coherent, and more effective at distinguishing topics than those of unmerged models. %R 10.18653/v1/2022.findings-acl.212 %U https://aclanthology.org/2022.findings-acl.212 %U https://doi.org/10.18653/v1/2022.findings-acl.212 %P 2696-2704