Abstract
Learning vector quantization (LVQ) as proposed by Kohonen is a simple and intuitive, though very successful prototype-based clustering algorithm. Generalized relevance LVQ (GRLVQ) constitutes a modification which obeys the dynamics of a gradient descent and allows an adaptive metric utilizing relevance factors for the input dimensions. As iterative algorithms with local learning rules, LVQ and modifications crucially depend on the initialization of the prototypes. They often fail for multimodal data. We propose a variant of GRLVQ which introduces ideas of the neural gas algorithm incorporating a global neighborhood coordination of the prototypes. The resulting learning algorithm, supervised relevance neural gas, is capable of learning highly multimodal data, whereby it shares the benefits of a gradient dynamics and an adaptive metric with GRLVQ.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
References
T. Graepel, M. Burger, and K. Obermayer. Self-organizing maps: generalizations and new optimization techniques. Neurocomputing, 20:173–190, 1998.
B. Hammer, T. Villmann. Estimating relevant input dimensions for self-organizing algorithms. In: N. Allison, H. Yin, L. Allinson, J. Slack (eds.), Advances in Self-Organizing Maps, pp. 173–180, Springer, 2001.
T. Kohonen. Self-Organizing Maps. Springer, 1997.
T. Martinetz, S. Berkovich, and K. Schulten. ‘Neural-gas’ network for vector quantization and its application to time-series prediction. IEEE TNN4(4):558–569, 1993.
G. Patané and Marco Russo. The enhanced LBG algorithm. Neural Networks 14:1219–1237, 2001.
A. S. Sato, K. Yamada. Generalized learning vector quantization. In G. Tesauro, D. Touretzky, and T. Leen, editors, Advances in NIPS, volume 7, pp. 423–429. MIT Press, 1995.
T. Villmann, R. Der, M. Herrmann, and T. M. Martinetz, Toplogy Preservation in Self-Organizing Feature Maps: Exact Definition and Precise Measurement, IEEE TNN 8(2):256–266, 1997.
N. Vlassis and A. Likas. A greedy algorithm for Gaussian mixture learning. Neural Processing Letters 15(1): 77–87, 2002.
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2002 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Hammer, B., Strickert, M., Villmann, T. (2002). Learning Vector Quantization for Multimodal Data. In: Dorronsoro, J.R. (eds) Artificial Neural Networks — ICANN 2002. ICANN 2002. Lecture Notes in Computer Science, vol 2415. Springer, Berlin, Heidelberg. https://doi.org/10.1007/3-540-46084-5_60
Download citation
DOI: https://doi.org/10.1007/3-540-46084-5_60
Published:
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-540-44074-1
Online ISBN: 978-3-540-46084-8
eBook Packages: Springer Book Archive