Abstract
A genetic algorithm (GA) method that evolves both the topology and training parameters of backpropagation-trained, fully-connected, feed-forward neural networks is presented. The GA uses a weak encoding scheme with real-valued alleles. One contribution of the proposed approach is to replace the needed but potentially slow evolution of final weights by the more efficient evolution of a single weight spread parameter used to set the initial weights only. In addition, the co-evolution of an input mask effects a form of automatic feature selection. Preliminary experiments suggest that the resulting system is able to produce networks that perform well under backpropagation.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
Similar content being viewed by others
References
Burdsall, B. and Giraud-Carrier, C. (1997). Evolving Fuzzy Prototypes for Efficient Data Clustering. In Proceedings of the Second International ICSC Symposium on Fuzzy Logic and Applications (ISFL ′97), 217–223.
Fahlman, S.E. (1988). An Empirical Study of Learning Speed in Backpropagation Networks. Technical Report CMU-CS-88-162, Carnegie Mellon University.
Fullmer, B. and Miikkulainen, R. (1992). Using Marker-Based Genetic Encoding of Neural Networks to Evolve Finite-State Behaviour. In Proceedings of the First European Conference on Artificial Life (ECAL′91), 255–262.
Goldberg, D.E. (1989). Genetic Algorithms in Search, Optimization and Machine Learning. Addison-Wesley Publishing Company.
Harp, S.A., Samad, T. and Guha, A. (1989). Towards the Genetic Synthesis of ANN. In Proceedings of the Third International Conference on Genetic Algorithms (ICGA′89), 360–369.
Holland, J. (1975). Adaptation in Natural and Artificial Systems. The University of Michigan Press, Ann Arbor, MI.
Jacobs, R.A. (1988). Increased Rates of Convergence Through Learning Rate Adaptation. Neural Networks, 1(4):295–307.
Koza, J.R. (1992). Genetic Programming: On the Programming of Computers by Means of Natural Selection. MIT Press.
Lock, D.F. (1998). Using Genetic Algorithms to Build, train and Optimize Neural Networks. MSc Thesis, Department of Computer Science, University of Bristol.
Merz, C.J. and Murphy, P.M. (1996). UCI Repository of Machine Learning Databases. Department of Information and Computer Science, University of California, Irvine.
Miller, G.F., Todd, P.M. and Hedge, S.U. (1991). Designing Neural Networks. Neural Networks, 4: 53–60.
Plaut, D.C. and Hinton, G.E. (1987). Learning Sets of Filters Using Backpropagation. Computer Speech and Language, 2:35–61.
Rumelhart, D.E., Hinton, G.E. and Williams, R.J. (1986). Learning Internal Representations by Error Propagation. In Rumelhart, D.E. and McClelland, J.L. (Eds.), Parallel Distributed Processing, Vol. 1,MTT Press.
Author information
Authors and Affiliations
Rights and permissions
Copyright information
© 1999 Springer-Verlag Wien
About this paper
Cite this paper
Lock, D., Giraud-Carrier, C. (1999). Evolutionary Programming of Near-Optimal Neural Networks. In: Artificial Neural Nets and Genetic Algorithms. Springer, Vienna. https://doi.org/10.1007/978-3-7091-6384-9_50
Download citation
DOI: https://doi.org/10.1007/978-3-7091-6384-9_50
Publisher Name: Springer, Vienna
Print ISBN: 978-3-211-83364-3
Online ISBN: 978-3-7091-6384-9
eBook Packages: Springer Book Archive