Skip to main content

Competitive Learning by Information Maximization: Eliminating Dead Neurons in Competitive Learning

  • Conference paper
  • First Online:
Artificial Neural Networks and Neural Information Processing — ICANN/ICONIP 2003 (ICANN 2003, ICONIP 2003)

Part of the book series: Lecture Notes in Computer Science ((LNCS,volume 2714))

Abstract

In this paper, we propose a new information theoretic competitive learning method. In realizing competition, neither the winner-take-all algorithm nor the lateral inhibition is used. Instead, the new method is based upon mutual information maximization between input patterns and competitive units. In maximizing mutual information, the entropy of competitive units is increased as much as possible. This means that all competitive units must equally be used in our framework. Thus, no under-utilized neurons (dead neurons) are generated. We applied our method to a simple artificial data problem and an actual road classification problem. In both cases, experimental results confirmed that the new method can produce the final solutions almost independently of initial conditions, and classification performance is significantly improved.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 39.99
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. R. Linsker, “Self-organization in a perceptual network,” Computer, vol. 21, pp. 105–117, 1988.

    Article  Google Scholar 

  2. R. Linsker, “How to generate ordered maps by maximizing the mutual information between input and output,” Neural Computation, vol. 1, pp. 402–411, 1989.

    Article  Google Scholar 

  3. R. Kamimura and S. Nakanishi, “Hidden information maximization for feature detection and rule discovery,” Network, vol. 6, pp. 577–622, 1995.

    Article  MATH  Google Scholar 

  4. R. Kamimura and S. Nakanishi, “Improving generalization performance by information minimization,” IEICE Transactions on Information and Systems, vol. E78-D, no. 2, pp. 163–173, 1995.

    Google Scholar 

  5. G. Deco, W. Finnof, and H. G. Zimmermann, “Unsupervised mutual information criterion for elimination of overtraining in supervised multiplayer networks,” Neural Computation, vol. 7, pp. 86–107, 1995.

    Article  Google Scholar 

  6. Y. Akiyama and T. Furuya, “An extension of the back-propagation learning which performs entropy maximization as well as error minimization,” Tech. Rep. NC91-6, IEICE Technical Report, 1991.

    Google Scholar 

  7. R. Kamimura, T. Kamimura, and T. R. Shultz, “Information theoretic competitive learning and linguistic rule acquistion,” Transactions of the Japanese Society for Artificial Intelligence, vol. 16, no. 2, pp. 287–298, 2001.

    Article  Google Scholar 

  8. R. Kamimura, T. Kamimura, and O. Uchida, “Flexible feature discovery and structural information,” Connection Science, vol. 13, no. 4, pp. 323–347, 2001.

    Article  Google Scholar 

  9. D. DeSieno, “Adding a conscience to competitive learning,” in Proceedings of IEEE International Conference on Neural Networks, (San Diego), pp. 117–124, IEEE, 1988.

    Google Scholar 

  10. S. C. Ahalt, A. K. Krishnamurthy, P. Chen, and D. E. Melton, “Competitive learning algorithms for vector quantization,” Neural Networks, vol. 3, pp. 277–290, 1990.

    Article  Google Scholar 

  11. L. Xu, “Rival penalized competitive learning for clustering analysis, RBF net, and curve detection,” IEEE Transaction on Neural Networks, vol. 4, no. 4, pp. 636–649, 1993.

    Article  Google Scholar 

  12. A. Luk and S. Lien, “Properties of the generalized lotto-type competitive learning,” in Proceedings of International conference on neural information processing, (San Mateo: CA), pp. 1180–1185, Morgan Kaufmann Publishers, 2000.

    Google Scholar 

  13. M. Marc and M. V. Hulle, Faithful representations and topographic maps. New York: John Wiley and Sons, Inc, 2000.

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2003 Springer-Verlag Berlin Heidelberg

About this paper

Cite this paper

Kamimura, R. (2003). Competitive Learning by Information Maximization: Eliminating Dead Neurons in Competitive Learning. In: Kaynak, O., Alpaydin, E., Oja, E., Xu, L. (eds) Artificial Neural Networks and Neural Information Processing — ICANN/ICONIP 2003. ICANN ICONIP 2003 2003. Lecture Notes in Computer Science, vol 2714. Springer, Berlin, Heidelberg. https://doi.org/10.1007/3-540-44989-2_13

Download citation

  • DOI: https://doi.org/10.1007/3-540-44989-2_13

  • Published:

  • Publisher Name: Springer, Berlin, Heidelberg

  • Print ISBN: 978-3-540-40408-8

  • Online ISBN: 978-3-540-44989-8

  • eBook Packages: Springer Book Archive

Publish with us

Policies and ethics