Skip to main content

Reorganising Artificial Neural Network Topologies

Complexifying Neural Networks by Reorganisation

  • Chapter
Advances in Electrical Engineering and Computational Science

Part of the book series: Lecture Notes in Electrical Engineering ((LNEE,volume 39))

  • 2296 Accesses

This chapter describes a novel way of complexifying artificial neural networks through topological reorganisation. The neural networks are reorganised to optimise their neural complexity, which is a measure of the information-theoretic complexity of the network. Complexification of neural networks here happens through rearranging connections, i.e. removing one or more connections and placing them elsewhere. The results verify that a structural reorganisation can help to increase the probability of discovering a neural network capable of adequately solving complex tasks. The networks and the methodology proposed are tested in a simulation of a mobile robot racing around a track.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 129.00
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 169.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info
Hardcover Book
USD 169.99
Price excludes VAT (USA)
  • Durable hardcover edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. A.S. Weigend, D.E. Rumelhart, and B.A. Huberman, Back-propagation, weight-elimination and time series prediction, Proceedings of the 1990 Summer School on Connectionist Models (1990).

    Google Scholar 

  2. J. Denker, D. Schwartz, B. Wittner, S. Solla, R. Howard, L. Jackel, and J. Hopfield, Large Automatic Learning, Rule Extraction and Generalization, Complex Systems 1(5): 877–922 (1987).

    MATH  MathSciNet  Google Scholar 

  3. S. Nolfi, and D. Floreano, Evolutionary Robotics; The Biology, Intelligence, and Technology of Self-Organizing Machines (Cambridge, MA: MIT Press 2000).

    Google Scholar 

  4. X. Yao, and Y. Liu, A New Evolutionary System for Evolving Artificial Neural Networks, IEEE Transactions on Neural Networks 8(3): 694–713 (1997).

    Article  MathSciNet  Google Scholar 

  5. X. Yao, Evolving Artificial Neural Networks, Proceedings of the IEEE 87(9) (1999).

    Google Scholar 

  6. R. Jacobs, and M. Jordan, Adaptive Mixtures of Local Experts, Neural Computation 3: 79–87 (1991).

    Article  Google Scholar 

  7. S.E. Fahlman, and C. Lebiere, The Cascade-Correlation Learning Architecture, Advances in Neural Information Processing Systems 2: 524–532 (1990).

    Google Scholar 

  8. P. Angeline, and J. Pollack, Evolutionary Module Acquisition, Proceedings of the Second Annual Conference on Evolutionary Programming (1993).

    Google Scholar 

  9. O. Sporns, and M. Lungarella, Evolving Coordinated Behaviours by Maximizing Informational Structure, Proceedings of the Tenth International Conference on Artificial Life (2006).

    Google Scholar 

  10. K.O. Stanley, and R. Miikkulainen, Continual Coevolution Through Complexification, Proceedings of the Genetic and Evolutionary Conference (2002).

    Google Scholar 

  11. G. Tononi, O. Sporns, and G.M. Edelman, A Measure for Brain Complexity: Relating Functional Segregation and Integration in the Nervous System, Proceedings of the National Academy of Science of USA (1994).

    Google Scholar 

  12. L.S. Yaeger, and O. Sporns, Evolution of Neural Structure and Complexity in a Computational Ecology, Proceedings of the Tenth International Conference on Artificial Life (2006).

    Google Scholar 

  13. R. Dawkins, Climbing Mount Improbable (Reprint by Penguin Books, London, England, 2006).

    Google Scholar 

  14. G.N. Martin, Human Neuropsychology (London: Prentice Hall, 1998, Reprinted 1999).

    Google Scholar 

  15. G. Edelman, Neural Darwinism — The Theory of Neuronal Group Selection (New York: Basic Books, 1989, Print by Oxford Press 1990).

    Google Scholar 

  16. C.E. Shannon, A Mathematical Theory of Communication, The Bell System Technical Journal 27: 379–423/623–656 (1948).

    MATH  MathSciNet  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2009 Springer Science+Business Media B.V

About this chapter

Cite this chapter

Jorgensen, T.D., Haynes, B., Norlund, C. (2009). Reorganising Artificial Neural Network Topologies. In: Ao, SI., Gelman, L. (eds) Advances in Electrical Engineering and Computational Science. Lecture Notes in Electrical Engineering, vol 39. Springer, Dordrecht. https://doi.org/10.1007/978-90-481-2311-7_35

Download citation

  • DOI: https://doi.org/10.1007/978-90-481-2311-7_35

  • Publisher Name: Springer, Dordrecht

  • Print ISBN: 978-90-481-2310-0

  • Online ISBN: 978-90-481-2311-7

  • eBook Packages: EngineeringEngineering (R0)

Publish with us

Policies and ethics