Abstract
In this paper, we introduce a novel method for aggregating multiple topic models to produce an aggregate model that contains topics with greater coherence than individual models. When generating a topic model a number of parameters must be specified. Depending on the parameters chosen the resulting topics can be very general or very specific. In this paper the process of aggregating multiple topic models generated using different parameters is investigated; the hypothesis being that combining the general and specific topics can increase topic coherence. The aggregate model is created using cosine similarity and Jensen-Shannon divergence to combine topics which are above a similarity threshold. The model is evaluated using evaluation methods to calculate the coherence of topics in the base models against those of the aggregated model. The results presented in this paper show that the aggregated model outperforms standard topic models at a statistically significant level in terms of topic coherence when evaluated against an external corpus.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Notes
- 1.
Available at: http://www.cs.princeton.edu/~blei/lda-c/.
References
Aletras, N., Stevenson, M.: Measuring the Similarity between automatically generated topics. In: Proceedings of the 14th Conference of the European Chapter of the Association for Computational Linguistics, pp. 22–27, Gothenburg (2014)
Blei, D.M., Griffiths, T.L., Jordan, M.I., Tenenbaum, J.B.: Hierarchical topic models and the nested Chinese restaurant process. Adv. Neural Inf. Process. Syst. 16, 17–24 (2004)
Blei, D.M., Ng, A., Jordan, M.I.: Latent Dirichlet allocation. J. Mach. Learn. Res. 3, 993–1022 (2003)
Chang, J., Gerrish, S., Wang, C., Boyd-Graber, J.L., Blei, D.M.: Reading tea leaves: how humans interpret topic models. In: Advances in Neural Information Processing Systems, pp. 288–296 (2009)
Deerwester, S., Dumais, S., Furnas, G., Landauer, T., Harshman, R.: Indexing by latent semantic analysis. J. Am. Soc. Inf. Sci. 41(6), 391–407 (1990)
Hall, D., Jurafsky, D., Manning, C.D.: Studying the history of ideas using topic models. In: Proceedings of the Conference on Empirical Methods in Natural Language Processing, pp. 363–371 (2008)
Hofmann, T.: The cluster-abstraction model: unsupervised learning of topic hierarchies from text data. IJCAI 99, 682–687 (1999)
Lau, J.H., Newman, D., Baldwin, T.: Machine reading tea leaves: automatically evaluating topic coherence and topic model quality. In: Proceedings of the 14th Conference of the European Chapter of the Association for Computational Linguistics, pp. 530–539, Gothenburg (2014)
Li, W., McCallum, A.: Pachinko allocation: DAG-structured mixture models of topic correlations. In: Proceedings of the 23rd International Conference on Machine Learning, pp. 577–584 (2006)
Mimno, D., Wallach, H.M., Talley, E., Leenders, M., McCallum, A.: Optimizing semantic coherence in topic models. In: Proceedings of the Conference on Empirical Methods in Natural Language Processing, pp. 262–272. Association for Computational Linguistics, july 2011
Newman, D., Lau, J.H., Grieser, K., Baldwin, T.: Automatic evaluation of topic coherence. In: Human Language Technologies: The 2010 Annual Conference of the North American Chapter of the ACL, pp. 100–108, Los Angeles (2010)
Quan, X., Kit, C., Ge, Y., Pan, S.J.: Short and sparse text topic modeling via self-aggregation. In: Proceedings of the 24th International Joint Conference on Artificial Intelligence, pp. 2270–2276 (2015)
Rider, A.K., Chawla, N.V.: An ensemble topic model for sharing healthcare data and predicting disease risk. In: Proceedings of the International Conference on Bioinformatics, Computational Biology and Biomedical Informatics, pp. 333–340 (2013)
Shen, Z., Luo, P., Yang, S., Shen, X.: Topic modeling ensembles. In: 2010 IEEE International Conference on Data Mining. pp. 1031–1036 (2010)
Tang, J., Zhang, M., Mei, Q.: One theme in all views: modeling consensus topics in multiple contexts. In: Proceedings of the 19th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, pp. 5–13 (2013)
Wallach, H.M.: Structured topic models for language. Ph.D. thesis, Universty of Cambridge (2008)
Wallach, H.M., Mimno, D., Mccallum, A.: Rethinking LDA: why priors matter. In: NIPS (2009)
Wallach, H.M., Murray, I., Salakhutdinov, R., Mimno, D.: Evaluation methods for topic models. In: Proceedings of the 26th Annual International Conference on Machine Learning - ICML 2009, New York, USA, pp. 1–8. ACM, New York, June 2009
Yan, X., Guo, J., Lan, Y., Cheng, X.: A biterm topic model for short texts. In: Proceedings of the 22nd International Conference on World Wide Web, pp. 1445–1456 (2013)
Yongliang, W., Qiao, G.: Multi-LDA hybrid topic model with boosting strategy and its application in text classification. In: 33rd Chinese Control Conference, pp. 4802–4806 (2014)
Yongliang, W., Qiao, G.: Modeling texts in semantic space and ensemble topic-models via boosting strategy. In: 34th Chinese Control Conference, pp. 3838–3843 (2015)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2016 Springer International Publishing AG
About this paper
Cite this paper
Blair, S.J., Bi, Y., Mulvenna, M.D. (2016). Increasing Topic Coherence by Aggregating Topic Models. In: Lehner, F., Fteimi, N. (eds) Knowledge Science, Engineering and Management. KSEM 2016. Lecture Notes in Computer Science(), vol 9983. Springer, Cham. https://doi.org/10.1007/978-3-319-47650-6_6
Download citation
DOI: https://doi.org/10.1007/978-3-319-47650-6_6
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-47649-0
Online ISBN: 978-3-319-47650-6
eBook Packages: Computer ScienceComputer Science (R0)