Abstract
In this article, we propose a new learning method called “self-enhancement learning.” In this method, targets for learning are not given from the outside, but they can be spontaneously created within a neural network. To realize the method, we consider a neural network with two different states, namely, an enhanced and a relaxed state. The enhanced state is one in which the network responds very selectively to input patterns, while in the relaxed state, the network responds almost equally to input patterns. The gap between the two states can be reduced by minimizing the Kullback–Leibler divergence between the two states with free energy. To demonstrate the effectiveness of this method, we applied self-enhancement learning to the self-organizing maps, or SOM, in which lateral interactions were added to an enhanced state. We applied the method to the well-known Iris, wine, housing and cancer machine learning database problems. In addition, we applied the method to real-life data, a student survey. Experimental results showed that the U-matrices obtained were similar to those produced by the conventional SOM. Class boundaries were made clearer in the housing and cancer data. For all the data, except for the cancer data, better performance could be obtained in terms of quantitative and topological errors. In addition, we could see that the trustworthiness and continuity, referring to the quality of neighborhood preservation, could be improved by the self-enhancement learning. Finally, we used modern dimensionality reduction methods and compared their results with those obtained by the self-enhancement learning. The results obtained by the self-enhancement were not superior to but comparable with those obtained by the modern dimensionality reduction methods.
Similar content being viewed by others
References
Ahalt SC, Krishnamurthy AK, Chen P, Melton DE (1990) Competitive learning algorithms for vector quantization. Neural Netw 3: 277–290
Bakker B, Heskes T (2003) Clustering ensembles of neural network models. Neural Netw 16: 261–269
Baluja S (1998) Probabilistic modeling for face orientation Probabilistic modeling for face orientation. In Kearns MS, Solla SA, Cohn DA (eds) Advances in neural information processing systems. MIT Press, Cambridge
Bauer H-U, Pawelzik K (1992) Quantifying the neighborhood preservation of self-organizing maps. IEEE Trans Neural Netw 3(4): 570–578
Benett K, Dermiriz A (1999) Semi-supervised support vector machines. In Advances in neural information processing systems, vol 11
Bishop CM (1995) Neural networks for pattern recognition. Oxford University Press, Oxford
Blum A, Chawla S (2001) Learning from labeled and unlabeled data using graph mincuts. In Proceedings of the 18th international conference on machine learning
Blum A, Mitchell T (1998) Combining labeled and unlabeled data with co-training. In Proceedings of the workshop on computational learning theory
Brefeld U, gartner T, Scheffer T, Wrobel S (2006) Efficient co-regularised least squares regression. In Proceedings of the 23rd international conference on machine learning
Chapell, OZ, Scholkopf, B (eds) (2005) Semi-supervised learning. MIT Press, Cambridge
Collins M, Singer Y (1999) Unsupervised models for named entity classification. In Proceedings of EMNLP/VLC-99
de Sa VR (1994) Unsupervised classification learning from cross-modal environmental structure. Dissertation, University of Rochester, Rochester
de Sa VR (1998) Category learning through multimodality sensing. Neural Comput 10: 1097–1117
de Sa VR (2004) Sensory modality segregation. In Thrun S, Saul L, Schoelkopf B (eds) Advanced in neural information processing systems, vol 16, pp 913–920
Frank A, Asuncion A (2010) UCI machine learning repository. University of California, Irvine
Fukushima K (1975a) Cognitron: a self-organizing multi-layered neural network. Biol Cybern 20: 121–136
Fukushima K (1975b) Neocognitron: a hierarchical neural network capable of visual pattern recognition. Biol Cybern 20: 121–136
Fukushima K (1983) Neocognitron: a neural network model for a mechanism of visual pattern recognition. IEEE Trans Syst Man Cybernet 13: 826–834
Goodhill GJ, Sejnowski TJ (1997) A unifying objective function for topographic mappings. Neural Comput 9: 1291–1303
Graepel T, Burger M, Obermayer K (1997) Phase transitions in stochastic self-organizing maps. Phys Rev E 56(4): 3876
Heskes T (2001) Self-organizing maps, vector quantization, and mixture modeling. IEEE Trans Neural Netw 12(6): 1299–1305
Hulle MMV (1997) The formation of topographic maps that maximize the average mutual information of the output responses to noiseless input signals. Neural Comput 9(3): 595–606
Joachims T (2000) Transductive inference for text classification using support vector machines. In Proceedings of the 20th international conference on machine learning
Kamimura R (2003a) Information-theoretic competitive learning with inverse Euclidean distance output units. Neural Process Lett 18: 163–184
Kamimura R (2003b) Teacher-directed learning: information-theoretic competitive learning in supervised multi-layered networks. Connect Sci 15: 117–140
Kamimura R (2003c) Progressive feature extraction by greedy network-growing algorithm. Complex Syst 14(2): 127–153
Kamimura R (2003d) Information theoretic competitive learning in self-adaptive multi-layered networks. Connect Sci 13(4): 323–347
Kamimura R (2006) Cooperative information maximization with Gauissian activation functions for self-organizing maps. IEEE Trans Neural Netw 17(4): 909–919
Kamimura R (2008a) Free energy-based competitive learning for mutual information maximization. In Proceedings of IEEE conference on systems, man, and cybernetics, pp 223–227
Kamimura R (2008b) Free energy-based competitive learning for self-organizing maps. In Proceedings of artificial intelligence and applications, pp 414–419
Kamimura R, Kamimura T (2000) Structural information and linguistic rule extraction. In: Proceedings of ICONIP-2000, pp 720– 726
Kamimura R, Kamimura T, Uchida O (2001) Flexible feature discovery and structural information control. Connect Sci 13(4): 323– 347
Kaski S, Nikkilä J, Kohonen T (1998) Methods for interpreting a self-organized map in data analysis. In Verleysen M (ed) Proceedings of ESANN’98. 6th European symposium on artificial neural networks, April 22–24. D-Facto, Bruges, Brussels, Belgium, pp 185–190
Kaski S, Nikkila J, Oja M, Venna J, Toronen P, Castren E (2003) Trustworthiness and metrics in visualizing similarity of gene expression. BMC Bioinf 4:48
Kiviluoto K (1996) Topology preservation in self-organizing maps. In: Proceedings of the IEEE international conference on neural networks, pp 294–299
Kohonen T (1988) Self-organization and associative memory. Springer-Verlag, New York
Kohonen T (1990) The self-organizing maps. Proc IEEE 78(9): 1464–1480
Kohonen T (1995) Self-Organizing Maps. Springer-Verlag, Berlin
Laaksonen J, Hurri J, Oja E (1998) Experiments with a self-supervised adaptive classification strategy in on-line recognition of isolated handwritten latin characters. In: Proceedings of the 6th workshop on frontiers of handwriting recognition, pp 475–484
Lanyon LJ, Denham SL (2004a) A biased competition computational model of spatial and object-based attention mediating active visual search. Neurocomputing 58(60): 655–662
Lanyon LJ, Denham SL (2004b) A model of active visual search with object-based attention. Neural Netw 17: 873–897
Lee JA, Verleysen M (2008) Quality assessment of nonlinear dimensionality reduction based on K-ary neighborhoods. In: JMLR: workshop and conference proceedings, vol 4, pp 21–35
Linsker R (1988) Self-organization in a perceptual network. Computer 21: 105–117
Linsker R (1989) How to generate ordered maps by maximizing the mutual information between input and output. Neural Comput 1: 402–411
Linsker R (1992) Local synaptic rules suffice to maximize mutual information in a linear network. Neural Comput 4: 691–702
Linsker R (2005) Improved local learning rule for information maximization and related applications. Neural Netw 18: 261–265
Luk A, Lien S (2000) Properties of the generalized lotto-type competitive learning. In: Proceedings of international conference on neural information processing. Morgan Kaufmann Publishers, San Mateo, CA, pp 1180–1185
Luttrell SP (1994) A Bayesian analysis of self-organising maps. Neural Comput 6(5): 767–794
Mitchell T (1999) The role of unlabeled data in supervised learning. In: Proceedings of the sixth international colloquium on cognitive science
Nigam K, Ghani R (2000) Analyzing the effectiveness and applicability of co-training. In: Nineth international conference on information and knowledge management, pp 86–93
Nigam K, McCalum AK, Thrun S, Mitchell T (2000) Text classificatiobn from labeled and unlabeled documents using EM. Mach Learn 39: 103–134
Oja M, Serber GO, Blomberg J, Kaski S (2005) Self-organizing map-based discovery and visualization of human endogenous retroviral sequence groups. Int J Neural Syst 15(3): 163–179
Pang B, Lee L (2004) A sentimental education: sentiment analysis using subjectivity summarization based on minimum cuts. In Proceedings of the association for computational linguistics, pp 271–278
Polzlbauer G (2004) Survey and comparison of quality measures for self-organizing maps. In: Proceedings of the fifth workshop on Data Analysis (WDA04), pp 67–82
Ridella S, Rovetta S, Zunino R (2001) K-winner machines for pattern classification. IEEE Trans Neural Netw 12(2): 371–385
Rilloff E, Wiebe J, Wilson T (2003) Learning subjective nouns using extraction pattern bootstrappping. In: Proceedings of the seventh conference on natural language laearning (CONLL2003)
Rose K, Gurewitz E, Fox GC (1990) Statistical mechanics and phase transition in clustering. Phys Rev Lett 65(8): 945–948
Rose K, Gurewitz E, Fox GC (1992) Vector quantization by deterministic annealing. IEEE Trans Inf Theory 38(4): 1249–1257
Rumelhart DE, Zipser D (1985) discovery by competitive learning”. Cognitive Science 9: 75–112
Sindhwani V, Niyogi P, Belkin M (2005) A co-regularization approach to semi-supervised learning with multiple views. In Proceedings of the 22nd ICML workshop on learning with multiple views
Ueda N, Nakano R (1995) Deterministic annealing variant of the EM algorithm. In Tesauro G, Touretzky D, Alspector J (eds) Advances in neural information processing systems, pp 545–552
Ueda N, Nakano R (1998) Deterministic annealing EM algorithm. Neural Netw 11: 271–282
Utsugi A (1997) Hyperparameter selection for self-organizing maps. Neural Comput 9(3): 623–635
Utsugi A (1998) Density estimation by mixture models with smoothing priors. Neural Comput 10: 2115–2135
van del Maaten LJP, Postma EO, van den Herik HJ (2009) Dimension reduction: a comparative overview. Tilburg University technical report, TiCC-TR 2009-005
Venna J (2007) Dimensionality reduction for visual exploration of similarity structures. Dissertation, Helsinki University of Technology, Helsinki
Venna J, Kaski S (2001) Neighborhood preservation in nonlinear projection methods: an experimental study. In: Lecture Notes in Computer Science, vol 2130, pp 485–491
Vesanto J (1999) SOM-based data visualization methods. Intel Data Anal 3: 111–126
Vesanto J, Alhoniemi E (2000) Clustering of the self-organizing map. IEEE Trans Neural Netw 11: 586–600
Vesanto J, Himberg J, Alhoniemi E, Parhankangas J (2000) SOM toolbox for matlab 5. Technical report A57. Helsinki University of Technology, Helsinki
Villmann T, Herrmann RDM, Martinez T (1997) Topology preservation in self-organizing feature maps: exact definition and measurment. IEEE Trans Neural Netw 8(2): 256–266
Xiong H, Swamy MNS, Ahmad MO (2004) Competitive splitting for codebook initialization. IEEE Signal Process Lett 11: 474–477
Xu L (1993) Rival penalized competitive learning for clustering analysis, RBF net, and curve detection. IEEE Trans Neural Netw 4(4): 636–649
Yarowsky D (1995) Unsupervised word sense disambiguation rivaling supervised methods. In: Proceedings of the 33rd annual meeting of the association for computational linguistics, pp 189–196
Yen JC, Guo JI, Chen HC (1998) A new k-winners-take-all neural networks and its array architecture. IEEE Trans Neural Netw 9(5): 901–912
Zhang YJ, Liu ZQ (2002) Self-splitting competitive learning: a new on-line clustering paradigm. IEEE Trans Neural Netw 13(2): 369–380
Zhu X (2005) Semi-supervised learning literature survey. Technical Report 1530. Department of Computer Sciences, University of Wisconsin-Madison, Madison
Author information
Authors and Affiliations
Corresponding author
Rights and permissions
About this article
Cite this article
Kamimura, R. Self-enhancement learning: target-creating learning and its application to self-organizing maps. Biol Cybern 104, 305–338 (2011). https://doi.org/10.1007/s00422-011-0434-x
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s00422-011-0434-x