Abstract
Researchers in the new field of “developmental robotics” propose to provide robots with so-called developmental programs. Similar to the development of human infants, robots might use those programs to interact with humans and their environment for extended periods of time, and become smarter autonomously. In this paper we show how a neural network model developed by neuroscientists can be used by an autonomous robot to learn by trial-and-error when considering rewards delivered at arbitrary times, as would be the case of developmental robots interacting with humans in the real world.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
Similar content being viewed by others
References
A. Billard. Learning motor skills by imitation: a biologically inspired robotic model. Cybernetics and Systems, 32(1-2), 2001 (in press).
K. Doya. What are the computations in the cerebellum, the basal ganglia, and the cerebral cortex. Neural Networks, 12:961–974, 1999.
K.J. Lang and G.E. Hinton. A time-delay neural network architecture for speech recognition. Technical Report CMU-DS-88-152, Dept. of Computer Science, Carnegie Mellon University, Pittsburgh, PA, December 1988.
F. Mondada, E. Franzi, and P. Ienne. Mobile robot miniaturization: A tool for investigating in control algorithms. In Proceedings of the Third International Symposium on Experimental Robotics, Kyoto, Japan, 1993.
A. Pérez-Uribe. Structure-Adaptable Digital Neural Networks, chapter 6.A Neurocontroller Architecture for Autonomous Robots, pages 95–116. Swiss Federal Institute of Technology-Lausanne, Ph.D Thesis 2052, 1999.
A. Pérez-Uribe. Using a time-delay actor-critic neural architecture with dopamine-like reinforcement signal for learning in autonomous robots. In S. Wermter, J. Austin, and D. Willshaw, editors, Emergent Neural Computational Architectures based on Neuroscience. Springer-Verlag, 2001 (to appear).
S. Schaal. Is imitation learning the route to humanoid robots?. Trends in Cognitive Sciences, 3(6):233–242, 1999.
W. Schultz, P. Dayan, and P. Read Montague. A Neural Substrate of Prediction and Reward. Science, 275:1593–1599, 14 March 1997.
M. Sipper, E. Sanchez, D. Mange, M. Tomassini, A. Pérez-Uribe, and A. Stauffer. A Phylogenetic, Ontogenetic, and Epigenetic View of Bio-Inspired Hardware Systems. IEEE Transactions on Evolutionary Computation, 1(1):83–97, April 1997.
R.E. Suri and W. Schultz. A Neural Network Model With Dopamine-Like Reinforcement Signal That Learns a Spatial Delayed Responde Task. Neuroscience, 91(3):871–890, 1999.
R.E. Suri and W. Schultz. Internal Model Reproduces Anticipatory Neural Activity. (available at http://www.snl.salk.edu/suri), July 1999.
R.S. Sutton and A.G. Barto. Reinforcement Learning: An Introduction. The MIT Press, 1998.
J. Weng, W.S. Hwang, Y. Zhang, C. Yang, and R. Smith. Developmental Humanoids: Humanoids that Develop Skills Automatically. In Proceedings the first IEEE-RAS International Conference on Humanoid Robots, Cambridge MA, September 7-8 2000.
J. Weng, J. McClelland, A. Pentland, O. Sporns, I. Stockman, M. Sur, and E. Thelen. Autonomous Mental Development by Robots and Animals. Science, 291(5504):599–600, 2000.
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2001 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Pérez-Uribe*, A., Courant, M. (2001). Learning to Predict Variable-Delay Rewards and Its Role in Autonomous Developmental Robotics. In: Mira, J., Prieto, A. (eds) Bio-Inspired Applications of Connectionism. IWANN 2001. Lecture Notes in Computer Science, vol 2085. Springer, Berlin, Heidelberg. https://doi.org/10.1007/3-540-45723-2_59
Download citation
DOI: https://doi.org/10.1007/3-540-45723-2_59
Published:
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-540-42237-2
Online ISBN: 978-3-540-45723-7
eBook Packages: Springer Book Archive