Abstract
Multi-layer perceptrons (MLPs) have been widely used in classification and regression task. How to improve the training speed of MLPs has been an interesting field of research. Instead of the classical method, we try to train MLPs by a MiniMin model which can ensure that the weights of the last layer are optimal at each step. Significant improvement on training speed has been made using our method for several big benchmark data sets.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
References
Rumelnart, D.E., Hinton, G.E., Williams, R.J.: Learning reprensentation of backprogation errors. Nature 223, 533–536 (1986)
Hornik, K., Stinchcombe, M., White, H.: Multilayer feedforward networks are univesal approximators. Neural Networks 2, 359–366 (1989)
Rumelhart, D.E., Hinton, G.E., Williams, R.J.: Learning Internal Representations by Error Propagation. In: Parallel Distributed Processing: Exploration in the Microstructure of Cognition, pp. 318–362 (1986)
Battiti, R., Masulli, F.: BFGS Optimization for faster and automated supervised learing. In: International Neural Network conference, pp. 757–760 (1990)
Hagan, M.T., Menhaj, M.: Training feedforward networks with the Marquardt algrithm. IEEE Transactions on Neural Networks 5, 989–993 (1994)
Moller, M.F.: A Scaled conjugate gradient algorithm for fast supervised learning. Neural Network 6, 525–533 (1993)
Demuth, H., Beale, M.: Neural network toolbox for use with MATLAB. The MatWorks Inc., Natick (1998)
Hagan, M.T., Demuth, H.B., Beale, M.H.: Neural Network Design. PWS Publishing, Boston (1996)
Boyd, S., Vandenberghe, L.: Convex Optimization. Cambridge University Press, Cambridge (2004)
Michie, D., Spiegelhalter, D.J., Taylor, C.C.: Machine Learning, Neural and Statistical Classification. Prentice Hall, Englewood Cliffs (1994)
Rasmussen, C.E., Neal, R.M., Hinton, C.E., Van Gamp, D., Revow, M., Ghahramani, Z., Kustra, R., Tibshirani, R.: The Deleve Manual (1996)
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2005 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Bo, L., Wang, L., Jiao, L. (2005). Training Multi-layer Perceptrons Using MiniMin Approach. In: Hao, Y., et al. Computational Intelligence and Security. CIS 2005. Lecture Notes in Computer Science(), vol 3801. Springer, Berlin, Heidelberg. https://doi.org/10.1007/11596448_135
Download citation
DOI: https://doi.org/10.1007/11596448_135
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-540-30818-8
Online ISBN: 978-3-540-31599-5
eBook Packages: Computer ScienceComputer Science (R0)