Skip to main content

Memetic Evolutionary Learning for Local Unit Networks

  • Conference paper
Advances in Neural Networks - ISNN 2010 (ISNN 2010)

Part of the book series: Lecture Notes in Computer Science ((LNTCS,volume 6063))

Included in the following conference series:

  • 1776 Accesses

Abstract

In this work we propose two hybrid algorithms combining evolutionary search with optimization algorithms. One algorithm memetically combines global evolution with gradient descent local search, while the other is a two-step procedure combining linear optimization with evolutionary search. It is shown that these algorithms typically produce smaller local unit networks with performance similar to theoretically sound but large regularization networks.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 84.99
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 109.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. Haykin, S.: Neural Networks: a comprehensive foundation, 2nd edn. Tom Robins (1999)

    Google Scholar 

  2. Vidnerova, P., Neruda, R.: Testing error estimates for regularization and radial function networks. In: Proceedings of the ISNN 2009. LNCS. Springer, Heidelberg (2009)

    Google Scholar 

  3. Tikhonov, A., Arsenin, V.: Solutions of Ill-posed Problems. W.H. Winston, Washington (1977)

    MATH  Google Scholar 

  4. Poggio, T., Girosi, F.: A theory of networks for approximation and learning. Technical report, Cambridge, MA, USA, A. I. Memo No. 1140, C.B.I.P. Paper No. 31 (1989)

    Google Scholar 

  5. Poggio, T., Smale, S.: The mathematics of learning: Dealing with data. Notices of the AMS 50, 536–544 (2003)

    MathSciNet  Google Scholar 

  6. Neruda, R., Kudová, P.: Learning methods for radial basis functions networks. Future Generation Computer Systems 21, 1131–1142 (2005)

    Article  Google Scholar 

  7. Yao, X.: Evolving artificial neural networks. Proceedings of the IEEE 9(87), 1423–1447 (1999)

    Google Scholar 

  8. Stanley, K.O., D’Ambrosio, D., Gauci, J.: A hypercube-based indirect encoding for evolving large-scale neural networks. Artificial Life 15(2) (2009)

    Google Scholar 

  9. Prechelt, L.: PROBEN1 – a set of benchmarks and benchmarking rules for neural network training algorithms. Technical Report 21/94, Universitaet Karlsruhe (September 1994)

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2010 Springer-Verlag Berlin Heidelberg

About this paper

Cite this paper

Neruda, R., Vidnerová, P. (2010). Memetic Evolutionary Learning for Local Unit Networks. In: Zhang, L., Lu, BL., Kwok, J. (eds) Advances in Neural Networks - ISNN 2010. ISNN 2010. Lecture Notes in Computer Science, vol 6063. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-13278-0_68

Download citation

  • DOI: https://doi.org/10.1007/978-3-642-13278-0_68

  • Publisher Name: Springer, Berlin, Heidelberg

  • Print ISBN: 978-3-642-13277-3

  • Online ISBN: 978-3-642-13278-0

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics