Abstract
Topic modeling is a popular method for discovering semantic information from textual data, with latent Dirichlet allocation (LDA) being a representative model. Recently, researchers have explored the use of variational autoencoders (VAE) to improve the performance of LDA. However, there remain two major limitations: (1) the Dirichlet prior is inadequate to extract precise semantic information in VAE-LDA models, as it introduces a trade-off between the topic quality and the sparsity of representations; (2) new variants of VAE-LDA models with auxiliary variables generally ignore the correlation between latent variables in the inference process due to the Mean-Field assumption. To address these issues, in this paper, we propose a Sparsity Reinforced and Non-Mean-Field Topic Model (SpareNTM) with a bank of auxiliary Bernoulli variables in the generative process of LDA to further model the sparsity of document representations. Thus individual documents are forced to focus on a subset of topics by a corresponding Bernoulli topic selector. Then, instead of applying the mean-field assumption for the posterior approximation, we take full advantage of VAE to realize a non-mean-field approximation, which succeeds in preserving the connection of latent variables. Experiment results on three datasets (20NewsGroup, Wikitext-103, and SearchSnippets) show that our model outperforms recent topic models in terms of both topic quality and sparsity.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Notes
- 1.
The code is available at https://github.com/Nazzcjy/SpareNTM.
- 2.
- 3.
The Gumbel(0, 1) distribution can be sampled using inverse transform sampling by drawing \(u\sim \text{ Uniform(0, } \text{1) }\) and computing \(g=-\log (-\log (u))\).
- 4.
- 5.
- 6.
- 7.
- 8.
- 9.
- 10.
Instead of using a sliding window, we consider a whole document to identify co-occurrence.
- 11.
References
Blei, D.M., Kucukelbir, A., McAuliffe, J.D.: Variational inference: a review for statisticians. J. Am. Stat. Assoc. 112(518), 859–877 (2017)
Blei, D.M., Ng, A.Y., Jordan, M.I.: Latent Dirichlet allocation. JMLR 3(Jan), 993–1022 (2003)
Burkhardt, S., Kramer, S.: Decoupling sparsity and smoothness in the Dirichlet variational autoencoder topic model. JMLR 20, 131:1–131:27 (2019)
Card, D., Tan, C., Smith, N.A.: Neural models for documents with metadata. In: Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics, ACL 2018, Melbourne, Australia, 15–20 July 2018, pp. 2031–2040. Association for Computational Linguistics (2018)
Dieng, A.B., Wang, C., Gao, J., Paisley, J.W.: TopiCRNN: a recurrent neural network with long-range semantic dependency. In: 5th International Conference on Learning Representations, ICLR 2017, Toulon, France, 24–26 April 2017, Conference Track Proceedings. OpenReview.net (2017)
Dieng, A.B., Ruiz, F.J.R., Blei, D.M.: Topic modeling in embedding spaces. Trans. Assoc. Comput. Linguist. 8, 439–453 (2020)
Drefs, J., Guiraud, E., Lücke, J.: Evolutionary variational optimization of generative models. J. Mach. Learn. Res. 23, 21:1–21:51 (2022)
Drefs, J., Guiraud, E., Panagiotou, F., Lücke, J.: Direct evolutionary optimization of variational autoencoders with binary latents. In: Amini, M.R., Canu, S., Fischer, A., Guns, T., Kralj Novak, P., Tsoumakas, G. (eds.) ECML PKDD 2022. LNCS, vol. 13715, pp. 357–372. Springer, Cham (2022). https://doi.org/10.1007/978-3-031-26409-2_22
Fallah, K., Rozell, C.J.: Variational sparse coding with learned thresholding. In: ICML. Proceedings of Machine Learning Research, vol. 162, pp. 6034–6058. PMLR (2022)
Feng, J., Zhang, Z., Ding, C., Rao, Y., Xie, H., Wang, F.L.: Context reinforced neural topic modeling over short texts. Inf. Sci. 607, 79–91 (2022)
Griffiths, T.L., Steyvers, M.: Finding scientific topics. Proc. Natl. Acad. Sci. 101(suppl_1), 5228–5235 (2004)
Jang, E., Gu, S., Poole, B.: Categorical reparameterization with gumbel-softmax. In: ICLR (2017)
Kingma, D.P., Welling, M.: Auto-encoding variational Bayes. In: ICLR (2014)
Lang, K.: Newsweeder: learning to filter netnews. In: Machine Learning, Proceedings of the Twelfth International Conference on Machine Learning, Tahoe City, California, USA, 9–12 July 1995, pp. 331–339 (1995). https://doi.org/10.1016/b978-1-55860-377-6.50048-7
Lau, J.H., Newman, D., Baldwin, T.: Machine reading tea leaves: automatically evaluating topic coherence and topic model quality. In: EACL, pp. 530–539 (2014)
Lin, T., Hu, Z., Guo, X.: Sparsemax and relaxed Wasserstein for topic sparsity. In: WSDM, pp. 141–149 (2019)
Merity, S., Xiong, C., Bradbury, J., Socher, R.: Pointer sentinel mixture models. In: 5th International Conference on Learning Representations, ICLR 2017, Toulon, France, 24–26 April 2017, Conference Track Proceedings (2017)
Miao, Y., Grefenstette, E., Blunsom, P.: Discovering discrete latent topics with neural variational inference. In: ICML, vol. 70, pp. 2410–2419 (2017)
Miao, Y., Yu, L., Blunsom, P.: Neural variational inference for text processing. In: ICML, vol. 48, pp. 1727–1736 (2016)
Mitchell, T.J., Beauchamp, J.J.: Bayesian variable selection in linear regression. J. Am. Stat. Assoc. 83(404), 1023–1032 (1988)
Naesseth, C.A., Ruiz, F.J.R., Linderman, S.W., Blei, D.M.: Reparameterization gradients through acceptance-rejection sampling algorithms. In: AISTATS. Proceedings of Machine Learning Research, vol. 54, pp. 489–498 (2017)
Nan, F., Ding, R., Nallapati, R., Xiang, B.: Topic modeling with Wasserstein autoencoders. In: Proceedings of the 57th Conference of the Association for Computational Linguistics, ACL 2019, Florence, Italy, 28 July–2 August 2019, pp. 6345–6381 (2019)
Nguyen, T., Luu, A.T.: Contrastive learning for neural topic model. In: Advances in Neural Information Processing Systems 34: Annual Conference on Neural Information Processing Systems 2021, NeurIPS 2021, 6–14 December 2021, Virtual, pp. 11974–11986 (2021)
Ning, X., et al.: Nonparametric topic modeling with neural inference. Neurocomputing 399, 296–306 (2020)
Peng, M., et al.: Neural sparse topical coding. In: ACL, pp. 2332–2340 (2018)
Phan, X.H., Nguyen, L., Horiguchi, S.: Learning to classify short and sparse text & web with hidden topics from large-scale data collections. In: Proceedings of the 17th International Conference on World Wide Web, WWW 2008, Beijing, China, 21–25 April 2008, pp. 91–100 (2008). https://doi.org/10.1145/1367497.1367510
Rezaee, M., Ferraro, F.: A discrete variational recurrent topic model without the reparametrization trick. In: Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M., Lin, H. (eds.) Advances in Neural Information Processing Systems 33: Annual Conference on Neural Information Processing Systems 2020, NeurIPS 2020, 6–12 December 2020, Virtual (2020)
Song, Z., Hu, Y., Verma, A., Buckeridge, D.L., Li, Y.: Automatic phenotyping by a seed-guided topic model. In: KDD, pp. 4713–4723. ACM (2022)
Srivastava, A., Sutton, C.: Autoencoding variational inference for topic models. In: ICLR (2017)
Srivastava, A., Sutton, C.: Variational inference in pachinko allocation machines. CoRR (2018)
Tian, R., Mao, Y., Zhang, R.: Learning VAE-LDA models with rounded reparameterization trick. In: EMNLP, pp. 1315–1325 (2020)
Tonolini, F., Jensen, B.S., Murray-Smith, R.: Variational sparse coding. In: UAI. Proceedings of Machine Learning Research, vol. 115, pp. 690–700. AUAI Press (2019)
Turner, R.E., Sahani, M.: Two problems with variational expectation maximisation for time-series models. In: Barber, D., Cemgil, T., Chiappa, S. (eds.) Bayesian Time Series Models, chap. 5, pp. 109–130. Cambridge University Press (2011)
Wang, C., Blei, D.M.: Decoupling sparsity and smoothness in the discrete hierarchical Dirichlet process. In: NeurIPS, pp. 1982–1989 (2009)
Wang, D., et al.: Representing mixtures of word embeddings with mixtures of topic embeddings. In: ICLR. OpenReview.net (2022)
Wang, R., et al.: Neural topic modeling with bidirectional adversarial training. In: Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, ACL 2020, Online, 5–10 July 2020, pp. 340–350 (2020)
Wang, R., Zhou, D., He, Y.: ATM: adversarial-neural topic model. Inf. Process. Manag. 56(6) (2019)
Williamson, S., Wang, C., Heller, K.A., Blei, D.M.: The IBP compound Dirichlet process and its application to focused topic modeling. In: ICML, pp. 1151–1158 (2010)
Wu, X., Li, C., Zhu, Y., Miao, Y.: Short text topic modeling with topic distribution quantization and negative sampling decoder. In: Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing, EMNLP, pp. 1772–1782 (2020)
Xu, J., Xu, B., Wang, P., Zheng, S., Tian, G., Zhao, J.: Self-taught convolutional neural networks for short text clustering. Neural Netw. 88, 22–31 (2017)
Zhao, H., Phung, D.Q., Huynh, V., Jin, Y., Du, L., Buntine, W.L.: Topic modelling meets deep neural networks: a survey. In: IJCAI, pp. 4713–4720 (2021)
Zhu, J., Xing, E.P.: Sparse topical coding. In: UAI, pp. 831–838 (2011)
Acknowledgment
This work was supported by the National Natural Science Foundation of China (61972261), Natural Science Foundation of Guangdong Province (2023A1515011667), Key Basic Research Foundation of Shenzhen (JCYJ20220818100205012), and Basic Research Foundations of Shenzhen (JCYJ20210324093609026).
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Ethics declarations
Ethical Statement
This paper doesn’t involve unethical tasks. All the experiment was done in public datasets.
Rights and permissions
Copyright information
© 2023 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this paper
Cite this paper
Chen, J., Wang, R., He, J., Li, M.J. (2023). Encouraging Sparsity in Neural Topic Modeling with Non-Mean-Field Inference. In: Koutra, D., Plant, C., Gomez Rodriguez, M., Baralis, E., Bonchi, F. (eds) Machine Learning and Knowledge Discovery in Databases: Research Track. ECML PKDD 2023. Lecture Notes in Computer Science(), vol 14172. Springer, Cham. https://doi.org/10.1007/978-3-031-43421-1_9
Download citation
DOI: https://doi.org/10.1007/978-3-031-43421-1_9
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-43420-4
Online ISBN: 978-3-031-43421-1
eBook Packages: Computer ScienceComputer Science (R0)