Abstract
This contribution shows that Gaussian Mixture Models can be considered generalizations of self-organizing maps. More precisely, we demonstrate that the training of self-organizing maps is an approximation to the training of Gaussian Mixture Models by gradient descent. As a consequence, the scores of a trained SOM can be treated as log-likelihoods of a GMM with tied, spherical covariance and used, e.g., for outlier detection, whereas sampling from trained SOMs is not well-defined. Furthermore, we outline how SGD-trained GMMs can be generalized to diagonal and more expressive covariance matrices and how this benefits typical data science applications such as outlier detection, sampling and generative classification. Source codes are available on the author’s web site or upon request.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
References
Gepperth, A.: Large-scale gradient-based training of mixture of factor analyzers. In: International Joint Conference on Neural Networks (IJCNN) (2022)
Gepperth, A., Pfülb, B.: Gradient-based training of Gaussian mixture models for high-dimensional streaming data. Neural Process. Lett. (2021, accepted)
Heskes, T.: Energy functions for self-organizing maps. In: Kohonen Maps, pp. 303–315. Elsevier (1999)
Heskes, T.: Self-organizing maps, vector quantization, and mixture modeling. IEEE Trans. Neural Netw. 12(6), 1299–1305 (2001)
Kohonen, T.: The self-organizing map. Proc. IEEE 78(9), 1464–1480 (1990)
LeCun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition. Proc. IEEE 86(11), 2278–2324 (1998)
Mureşan, H., Oltean, M.: Fruit recognition from images using deep learning. Acta Universitatis Sapientiae Informatica 10(1), 26–42 (2018). https://doi.org/10.2478/ausi-2018-0002
Richardson, E., Weiss, Y.: On GANs and GMMs. In: Advances in Neural Information Processing Systems (NeurIPS), pp. 5847–5858 (2018)
Verbeek, J.J., Vlassis, N., Kröse, B.J.: Self-organizing mixture models. Neurocomputing 63, 99–123 (2005)
Xiao, H., Rasul, K., Vollgraf, R.: Fashion-MNIST: a Novel Image Dataset for Benchmarking Machine Learning Algorithms, pp. 1–6 (2017). http://arxiv.org/abs/1708.07747
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2024 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this paper
Cite this paper
Gepperth, A. (2024). Generalizing Self-organizing Maps: Large-Scale Training of GMMs and Applications in Data Science. In: Villmann, T., Kaden, M., Geweniger, T., Schleif, FM. (eds) Advances in Self-Organizing Maps, Learning Vector Quantization, Interpretable Machine Learning, and Beyond. WSOM+ 2024. Lecture Notes in Networks and Systems, vol 1087. Springer, Cham. https://doi.org/10.1007/978-3-031-67159-3_7
Download citation
DOI: https://doi.org/10.1007/978-3-031-67159-3_7
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-67158-6
Online ISBN: 978-3-031-67159-3
eBook Packages: Intelligent Technologies and RoboticsIntelligent Technologies and Robotics (R0)