%0 Conference Proceedings %T Neural Mixed Counting Models for Dispersed Topic Discovery %A Wu, Jiemin %A Rao, Yanghui %A Zhang, Zusheng %A Xie, Haoran %A Li, Qing %A Wang, Fu Lee %A Chen, Ziye %Y Jurafsky, Dan %Y Chai, Joyce %Y Schluter, Natalie %Y Tetreault, Joel %S Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics %D 2020 %8 July %I Association for Computational Linguistics %C Online %F wu-etal-2020-neural %X Mixed counting models that use the negative binomial distribution as the prior can well model over-dispersed and hierarchically dependent random variables; thus they have attracted much attention in mining dispersed document topics. However, the existing parameter inference method like Monte Carlo sampling is quite time-consuming. In this paper, we propose two efficient neural mixed counting models, i.e., the Negative Binomial-Neural Topic Model (NB-NTM) and the Gamma Negative Binomial-Neural Topic Model (GNB-NTM) for dispersed topic discovery. Neural variational inference algorithms are developed to infer model parameters by using the reparameterization of Gamma distribution and the Gaussian approximation of Poisson distribution. Experiments on real-world datasets indicate that our models outperform state-of-the-art baseline models in terms of perplexity and topic coherence. The results also validate that both NB-NTM and GNB-NTM can produce explainable intermediate variables by generating dispersed proportions of document topics. %R 10.18653/v1/2020.acl-main.548 %U https://aclanthology.org/2020.acl-main.548 %U https://doi.org/10.18653/v1/2020.acl-main.548 %P 6159-6169