Skip to main content
Log in

Self-enhancement learning: target-creating learning and its application to self-organizing maps

  • Original Paper
  • Published:
Biological Cybernetics Aims and scope Submit manuscript

Abstract

In this article, we propose a new learning method called “self-enhancement learning.” In this method, targets for learning are not given from the outside, but they can be spontaneously created within a neural network. To realize the method, we consider a neural network with two different states, namely, an enhanced and a relaxed state. The enhanced state is one in which the network responds very selectively to input patterns, while in the relaxed state, the network responds almost equally to input patterns. The gap between the two states can be reduced by minimizing the Kullback–Leibler divergence between the two states with free energy. To demonstrate the effectiveness of this method, we applied self-enhancement learning to the self-organizing maps, or SOM, in which lateral interactions were added to an enhanced state. We applied the method to the well-known Iris, wine, housing and cancer machine learning database problems. In addition, we applied the method to real-life data, a student survey. Experimental results showed that the U-matrices obtained were similar to those produced by the conventional SOM. Class boundaries were made clearer in the housing and cancer data. For all the data, except for the cancer data, better performance could be obtained in terms of quantitative and topological errors. In addition, we could see that the trustworthiness and continuity, referring to the quality of neighborhood preservation, could be improved by the self-enhancement learning. Finally, we used modern dimensionality reduction methods and compared their results with those obtained by the self-enhancement learning. The results obtained by the self-enhancement were not superior to but comparable with those obtained by the modern dimensionality reduction methods.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Subscribe and save

Springer+ Basic
$34.99 /Month
  • Get 10 units per month
  • Download Article/Chapter or eBook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Similar content being viewed by others

References

  • Ahalt SC, Krishnamurthy AK, Chen P, Melton DE (1990) Competitive learning algorithms for vector quantization. Neural Netw 3: 277–290

    Article  Google Scholar 

  • Bakker B, Heskes T (2003) Clustering ensembles of neural network models. Neural Netw 16: 261–269

    Article  PubMed  Google Scholar 

  • Baluja S (1998) Probabilistic modeling for face orientation Probabilistic modeling for face orientation. In Kearns MS, Solla SA, Cohn DA (eds) Advances in neural information processing systems. MIT Press, Cambridge

  • Bauer H-U, Pawelzik K (1992) Quantifying the neighborhood preservation of self-organizing maps. IEEE Trans Neural Netw 3(4): 570–578

    Article  PubMed  CAS  Google Scholar 

  • Benett K, Dermiriz A (1999) Semi-supervised support vector machines. In Advances in neural information processing systems, vol 11

  • Bishop CM (1995) Neural networks for pattern recognition. Oxford University Press, Oxford

    Google Scholar 

  • Blum A, Chawla S (2001) Learning from labeled and unlabeled data using graph mincuts. In Proceedings of the 18th international conference on machine learning

  • Blum A, Mitchell T (1998) Combining labeled and unlabeled data with co-training. In Proceedings of the workshop on computational learning theory

  • Brefeld U, gartner T, Scheffer T, Wrobel S (2006) Efficient co-regularised least squares regression. In Proceedings of the 23rd international conference on machine learning

  • Chapell, OZ, Scholkopf, B (eds) (2005) Semi-supervised learning. MIT Press, Cambridge

    Google Scholar 

  • Collins M, Singer Y (1999) Unsupervised models for named entity classification. In Proceedings of EMNLP/VLC-99

  • de Sa VR (1994) Unsupervised classification learning from cross-modal environmental structure. Dissertation, University of Rochester, Rochester

  • de Sa VR (1998) Category learning through multimodality sensing. Neural Comput 10: 1097–1117

    Article  PubMed  CAS  Google Scholar 

  • de Sa VR (2004) Sensory modality segregation. In Thrun S, Saul L, Schoelkopf B (eds) Advanced in neural information processing systems, vol 16, pp 913–920

  • Frank A, Asuncion A (2010) UCI machine learning repository. University of California, Irvine

    Google Scholar 

  • Fukushima K (1975a) Cognitron: a self-organizing multi-layered neural network. Biol Cybern 20: 121–136

    Article  PubMed  CAS  Google Scholar 

  • Fukushima K (1975b) Neocognitron: a hierarchical neural network capable of visual pattern recognition. Biol Cybern 20: 121–136

    Article  PubMed  CAS  Google Scholar 

  • Fukushima K (1983) Neocognitron: a neural network model for a mechanism of visual pattern recognition. IEEE Trans Syst Man Cybernet 13: 826–834

    Google Scholar 

  • Goodhill GJ, Sejnowski TJ (1997) A unifying objective function for topographic mappings. Neural Comput 9: 1291–1303

    Article  Google Scholar 

  • Graepel T, Burger M, Obermayer K (1997) Phase transitions in stochastic self-organizing maps. Phys Rev E 56(4): 3876

    Article  CAS  Google Scholar 

  • Heskes T (2001) Self-organizing maps, vector quantization, and mixture modeling. IEEE Trans Neural Netw 12(6): 1299–1305

    Article  PubMed  CAS  Google Scholar 

  • Hulle MMV (1997) The formation of topographic maps that maximize the average mutual information of the output responses to noiseless input signals. Neural Comput 9(3): 595–606

    Article  Google Scholar 

  • Joachims T (2000) Transductive inference for text classification using support vector machines. In Proceedings of the 20th international conference on machine learning

  • Kamimura R (2003a) Information-theoretic competitive learning with inverse Euclidean distance output units. Neural Process Lett 18: 163–184

    Article  Google Scholar 

  • Kamimura R (2003b) Teacher-directed learning: information-theoretic competitive learning in supervised multi-layered networks. Connect Sci 15: 117–140

    Article  Google Scholar 

  • Kamimura R (2003c) Progressive feature extraction by greedy network-growing algorithm. Complex Syst 14(2): 127–153

    Google Scholar 

  • Kamimura R (2003d) Information theoretic competitive learning in self-adaptive multi-layered networks. Connect Sci 13(4): 323–347

    Article  Google Scholar 

  • Kamimura R (2006) Cooperative information maximization with Gauissian activation functions for self-organizing maps. IEEE Trans Neural Netw 17(4): 909–919

    Article  PubMed  Google Scholar 

  • Kamimura R (2008a) Free energy-based competitive learning for mutual information maximization. In Proceedings of IEEE conference on systems, man, and cybernetics, pp 223–227

  • Kamimura R (2008b) Free energy-based competitive learning for self-organizing maps. In Proceedings of artificial intelligence and applications, pp 414–419

  • Kamimura R, Kamimura T (2000) Structural information and linguistic rule extraction. In: Proceedings of ICONIP-2000, pp 720– 726

  • Kamimura R, Kamimura T, Uchida O (2001) Flexible feature discovery and structural information control. Connect Sci 13(4): 323– 347

    Article  Google Scholar 

  • Kaski S, Nikkilä J, Kohonen T (1998) Methods for interpreting a self-organized map in data analysis. In Verleysen M (ed) Proceedings of ESANN’98. 6th European symposium on artificial neural networks, April 22–24. D-Facto, Bruges, Brussels, Belgium, pp 185–190

  • Kaski S, Nikkila J, Oja M, Venna J, Toronen P, Castren E (2003) Trustworthiness and metrics in visualizing similarity of gene expression. BMC Bioinf 4:48

    Google Scholar 

  • Kiviluoto K (1996) Topology preservation in self-organizing maps. In: Proceedings of the IEEE international conference on neural networks, pp 294–299

  • Kohonen T (1988) Self-organization and associative memory. Springer-Verlag, New York

    Google Scholar 

  • Kohonen T (1990) The self-organizing maps. Proc IEEE 78(9): 1464–1480

    Article  Google Scholar 

  • Kohonen T (1995) Self-Organizing Maps. Springer-Verlag, Berlin

    Google Scholar 

  • Laaksonen J, Hurri J, Oja E (1998) Experiments with a self-supervised adaptive classification strategy in on-line recognition of isolated handwritten latin characters. In: Proceedings of the 6th workshop on frontiers of handwriting recognition, pp 475–484

  • Lanyon LJ, Denham SL (2004a) A biased competition computational model of spatial and object-based attention mediating active visual search. Neurocomputing 58(60): 655–662

    Article  Google Scholar 

  • Lanyon LJ, Denham SL (2004b) A model of active visual search with object-based attention. Neural Netw 17: 873–897

    Article  PubMed  Google Scholar 

  • Lee JA, Verleysen M (2008) Quality assessment of nonlinear dimensionality reduction based on K-ary neighborhoods. In: JMLR: workshop and conference proceedings, vol 4, pp 21–35

  • Linsker R (1988) Self-organization in a perceptual network. Computer 21: 105–117

    Article  Google Scholar 

  • Linsker R (1989) How to generate ordered maps by maximizing the mutual information between input and output. Neural Comput 1: 402–411

    Article  Google Scholar 

  • Linsker R (1992) Local synaptic rules suffice to maximize mutual information in a linear network. Neural Comput 4: 691–702

    Article  Google Scholar 

  • Linsker R (2005) Improved local learning rule for information maximization and related applications. Neural Netw 18: 261–265

    Article  Google Scholar 

  • Luk A, Lien S (2000) Properties of the generalized lotto-type competitive learning. In: Proceedings of international conference on neural information processing. Morgan Kaufmann Publishers, San Mateo, CA, pp 1180–1185

  • Luttrell SP (1994) A Bayesian analysis of self-organising maps. Neural Comput 6(5): 767–794

    Article  Google Scholar 

  • Mitchell T (1999) The role of unlabeled data in supervised learning. In: Proceedings of the sixth international colloquium on cognitive science

  • Nigam K, Ghani R (2000) Analyzing the effectiveness and applicability of co-training. In: Nineth international conference on information and knowledge management, pp 86–93

  • Nigam K, McCalum AK, Thrun S, Mitchell T (2000) Text classificatiobn from labeled and unlabeled documents using EM. Mach Learn 39: 103–134

    Article  Google Scholar 

  • Oja M, Serber GO, Blomberg J, Kaski S (2005) Self-organizing map-based discovery and visualization of human endogenous retroviral sequence groups. Int J Neural Syst 15(3): 163–179

    Article  PubMed  Google Scholar 

  • Pang B, Lee L (2004) A sentimental education: sentiment analysis using subjectivity summarization based on minimum cuts. In Proceedings of the association for computational linguistics, pp 271–278

  • Polzlbauer G (2004) Survey and comparison of quality measures for self-organizing maps. In: Proceedings of the fifth workshop on Data Analysis (WDA04), pp 67–82

  • Ridella S, Rovetta S, Zunino R (2001) K-winner machines for pattern classification. IEEE Trans Neural Netw 12(2): 371–385

    Article  PubMed  CAS  Google Scholar 

  • Rilloff E, Wiebe J, Wilson T (2003) Learning subjective nouns using extraction pattern bootstrappping. In: Proceedings of the seventh conference on natural language laearning (CONLL2003)

  • Rose K, Gurewitz E, Fox GC (1990) Statistical mechanics and phase transition in clustering. Phys Rev Lett 65(8): 945–948

    Article  PubMed  Google Scholar 

  • Rose K, Gurewitz E, Fox GC (1992) Vector quantization by deterministic annealing. IEEE Trans Inf Theory 38(4): 1249–1257

    Article  Google Scholar 

  • Rumelhart DE, Zipser D (1985) discovery by competitive learning”. Cognitive Science 9: 75–112

    Article  Google Scholar 

  • Sindhwani V, Niyogi P, Belkin M (2005) A co-regularization approach to semi-supervised learning with multiple views. In Proceedings of the 22nd ICML workshop on learning with multiple views

  • Ueda N, Nakano R (1995) Deterministic annealing variant of the EM algorithm. In Tesauro G, Touretzky D, Alspector J (eds) Advances in neural information processing systems, pp 545–552

  • Ueda N, Nakano R (1998) Deterministic annealing EM algorithm. Neural Netw 11: 271–282

    Article  PubMed  CAS  Google Scholar 

  • Utsugi A (1997) Hyperparameter selection for self-organizing maps. Neural Comput 9(3): 623–635

    Article  Google Scholar 

  • Utsugi A (1998) Density estimation by mixture models with smoothing priors. Neural Comput 10: 2115–2135

    Article  PubMed  Google Scholar 

  • van del Maaten LJP, Postma EO, van den Herik HJ (2009) Dimension reduction: a comparative overview. Tilburg University technical report, TiCC-TR 2009-005

  • Venna J (2007) Dimensionality reduction for visual exploration of similarity structures. Dissertation, Helsinki University of Technology, Helsinki

  • Venna J, Kaski S (2001) Neighborhood preservation in nonlinear projection methods: an experimental study. In: Lecture Notes in Computer Science, vol 2130, pp 485–491

  • Vesanto J (1999) SOM-based data visualization methods. Intel Data Anal 3: 111–126

    Article  Google Scholar 

  • Vesanto J, Alhoniemi E (2000) Clustering of the self-organizing map. IEEE Trans Neural Netw 11: 586–600

    Article  PubMed  CAS  Google Scholar 

  • Vesanto J, Himberg J, Alhoniemi E, Parhankangas J (2000) SOM toolbox for matlab 5. Technical report A57. Helsinki University of Technology, Helsinki

  • Villmann T, Herrmann RDM, Martinez T (1997) Topology preservation in self-organizing feature maps: exact definition and measurment. IEEE Trans Neural Netw 8(2): 256–266

    Article  PubMed  CAS  Google Scholar 

  • Xiong H, Swamy MNS, Ahmad MO (2004) Competitive splitting for codebook initialization. IEEE Signal Process Lett 11: 474–477

    Article  Google Scholar 

  • Xu L (1993) Rival penalized competitive learning for clustering analysis, RBF net, and curve detection. IEEE Trans Neural Netw 4(4): 636–649

    Article  PubMed  CAS  Google Scholar 

  • Yarowsky D (1995) Unsupervised word sense disambiguation rivaling supervised methods. In: Proceedings of the 33rd annual meeting of the association for computational linguistics, pp 189–196

  • Yen JC, Guo JI, Chen HC (1998) A new k-winners-take-all neural networks and its array architecture. IEEE Trans Neural Netw 9(5): 901–912

    Article  PubMed  CAS  Google Scholar 

  • Zhang YJ, Liu ZQ (2002) Self-splitting competitive learning: a new on-line clustering paradigm. IEEE Trans Neural Netw 13(2): 369–380

    Article  PubMed  Google Scholar 

  • Zhu X (2005) Semi-supervised learning literature survey. Technical Report 1530. Department of Computer Sciences, University of Wisconsin-Madison, Madison

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Ryotaro Kamimura.

Rights and permissions

Reprints and permissions

About this article

Cite this article

Kamimura, R. Self-enhancement learning: target-creating learning and its application to self-organizing maps. Biol Cybern 104, 305–338 (2011). https://doi.org/10.1007/s00422-011-0434-x

Download citation

  • Received:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s00422-011-0434-x

Keywords