Skip to main content
Log in

Towards the Optimal Learning Rate for Backpropagation

  • Published:
Neural Processing Letters Aims and scope Submit manuscript

Abstract

A backpropagation learning algorithm for feedforward neural networks withan adaptive learning rate is derived. The algorithm is based uponminimising the instantaneous output error and does not include anysimplifications encountered in the corresponding Least Mean Square (LMS)algorithms for linear adaptive filters. The backpropagation algorithmwith an adaptive learning rate, which is derived based upon the Taylorseries expansion of the instantaneous output error, is shown to exhibitbehaviour similar to that of the Normalised LMS (NLMS) algorithm. Indeed,the derived optimal adaptive learning rate of a neural network trainedby backpropagation degenerates to the learning rate of the NLMS for a linear activation function of a neuron. By continuity, the optimal adaptive learning rate for neural networks imposes additional stabilisationeffects to the traditional backpropagation learning algorithm.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Similar content being viewed by others

References

  1. Ljung, L. and Soderstrom, T.: Theory and Practice of Recursive Identification, MIT Press, Cambridge, MA.,1983.

    Google Scholar 

  2. Treichler, J. R., Johnson, Jr., C. R. and Larimore, M. G.: Theory and Design of Adaptive Filters, John Wiley & Sons, New York, 1987.

    Google Scholar 

  3. Kwong, R. H. and Johnston, E. W.: A variable step size LMS algorithm, IEEE Transactions on Signal Processing 40(7) (1992), 1633-1641.

    Google Scholar 

  4. Evans, J. B., Xue, P. and Liu, B.: Analysis and implementation of variable step size adaptive algorithms, IEEE Transactions on Signal Processing 41(8) (1993), 2517-2535.

    Google Scholar 

  5. Mathews, V. J. and Xie, Z.: A stochastic gradient adaptive filter with gradient adaptive step size, IEEE Transactions on Signal Processing 41(6) (1993), 2075-2087.

    Google Scholar 

  6. Shan, T. J. and Kailaith, T.: Adaptive algorithms with an automatic gain control feature, IEEE Transactions on Acoustics, Speech and Signal Processing 35(1) (1988), 122-127.

    Google Scholar 

  7. Aboulnasr, T. and Mayyas, K.: A robust variable step-size LMS-type algorithm: Analysis and simulations, IEEE Transactions on Signal Processing 45(3) (1997), 631-639.

    Google Scholar 

  8. Haykin, S.: Adaptive Filter Theory, Prentice-Hall, 3d ed., Englewood Cliffs, NJ, 1996.

    Google Scholar 

  9. Jacobs, R. A.: Increased rates of convergence through learning rate adaptation, Neural Networks 1 (1988), 295-307.

    Google Scholar 

  10. Haykin, S.: Neural Networks-A Comprehensive Foundation, Prentice Hall, Englewood Cliffs, NJ, 1994.

    Google Scholar 

  11. Douglas, S. C. and Cichocki, A.: On-line step-size selection for training of adaptive systems, IEEE Signal Processing Magazine 14(6) (1997), 45-46.

    Google Scholar 

  12. Mandic, D. P. and Chambers, J. A.: A posteriori real time recurrent learning schemes for a recurrent neural network based non-linear predictor. IEE Proceedings-Vision, Image and Signal Processing 145(6) (1998), 365-370.

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Rights and permissions

Reprints and permissions

About this article

Cite this article

Mandic, D.P., Chambers, J.A. Towards the Optimal Learning Rate for Backpropagation. Neural Processing Letters 11, 1–5 (2000). https://doi.org/10.1023/A:1009686825582

Download citation

  • Issue Date:

  • DOI: https://doi.org/10.1023/A:1009686825582

Navigation