Skip to main content

An Extremely Simple Reinforcement Learning Rule for Neural Networks

  • Conference paper
Advances in Neural Networks – ISNN 2007 (ISNN 2007)

Part of the book series: Lecture Notes in Computer Science ((LNTCS,volume 4491))

Included in the following conference series:

  • 1361 Accesses

Abstract

In this paper we derive a simple reinforcement learning rule based on a more general form of REINFORCE formulation. We test our new rule on both classification and reinforcement problems. The results have shown that although this simple learning rule has a high probability of being stuck in local optimum for the case of classification tasks, it is able to solve some global reinforcement problems (e.g. the cart-pole balancing problem) directly in the continuous space.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 129.00
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 169.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. Hertz, J., Palmer, R.G., Krogh, A.S.: Introduction to the theory of neural computation. Addison-Wesley Pub. Co., Redwood City (1991)

    Google Scholar 

  2. Sutton, R.S., Barto, A.G.: Reinforcement learning: An introduction. MIT Press, Cambridge (1998)

    Google Scholar 

  3. Williams, R.J.: Simple statistical gradient-following algorithms for connectionist reinforcement learning. Machine Learning 8, 229–256 (1992)

    MATH  Google Scholar 

  4. Ma, X., Likharev, K.K.: Global reinforcement learning in neural networks with stochastic synapses. In: Proc. of WCCI/IJCNN’06, pp. 47–53 (2006)

    Google Scholar 

  5. Ma, X., Likharev, K.K.: Global reinforcement learning in neural networks. To be published in IEEE Tran. on Neural Networks (2007)

    Google Scholar 

  6. Türel, Ö., Lee, J.H., Ma, X., Likharev, K.K.: Neuromorphic architectures for nanoelectronic circuits. Int. J. Circ. Theory App. 32(5), 277–302 (2004)

    Article  Google Scholar 

  7. Baxter, J., Bartlett, P.L.: Infinite-horizon policy-gradient estimation. Journal of Artificial Intelligence Research 15, 319–350 (2001)

    MathSciNet  MATH  Google Scholar 

  8. Barto, A.G., Sutton, R.S., Anderson, C.W.: Neuronlike adaptive elements that can solve difficult learning control problems. IEEE Trans. Syst., Man, Cybern. 13, 834–846 (1983)

    Article  Google Scholar 

  9. Albus, J.S.: A new approach to manipulator conrol: the cerebellar model articulation controller (CMAC). Trans. of ASME Journal of Dynamic Systems, Measurements, and Control 97(3), 220–227 (1975)

    Article  MATH  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2007 Springer-Verlag Berlin Heidelberg

About this paper

Cite this paper

Ma, X. (2007). An Extremely Simple Reinforcement Learning Rule for Neural Networks. In: Liu, D., Fei, S., Hou, ZG., Zhang, H., Sun, C. (eds) Advances in Neural Networks – ISNN 2007. ISNN 2007. Lecture Notes in Computer Science, vol 4491. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-540-72383-7_51

Download citation

  • DOI: https://doi.org/10.1007/978-3-540-72383-7_51

  • Publisher Name: Springer, Berlin, Heidelberg

  • Print ISBN: 978-3-540-72382-0

  • Online ISBN: 978-3-540-72383-7

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics