Skip to main content
Log in

Mobile Robot Learning by Self-Observation

  • Published:
Autonomous Robots Aims and scope Submit manuscript

Abstract

The aim of the project described in this paper was to investigate robot learning at a most fundamental level. The project focused on the transition between organisms with innate behaviors and organisms that have the most rudimentary capability of learning through their personal interaction with their environment. It was assumed that the innate behaviors gave basic survival competence but no learning ability. By observing the interaction between their innate behaviors and the organism's environment it was reasoned that the organism should be able to learn how to modify its actions in a way that improves its performance. If a learning system is given more information than it requires then, when it is successful, it is difficult to say which pieces of information contribute to the success. For this reason the information available to the learning system was kept to an absolute minimum. In order to provide a practical test of the learning scheme developed in this project, the robot environment EDEN was constructed. Within EDEN a robot's actions influence its internal ‘energy’ reserves. The environment incorporates sources of energy, and it also involves situations that use additional energy or reduce energy consumption. A successful learning scheme was developed purely based on the recorded history of the robot's interactions with its environment and the knowledge that the robot's innate behavior was reactive. This learning scheme allowed the robot to improve its energy management by exhibiting classical conditioning and a restricted form of operant conditioning.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Similar content being viewed by others

References

  • Bakker, B., Linaker, F., and Schmidhber, J. 2002. Reinforcement learning in partially observable robot domains using unsupervised event extraction. In Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems, Lausanne, Switzerland, vol. 1, pp. 938–943.

    Google Scholar 

  • Brooks, R.A. 1989. A robot that walks: emergent behaviors from a carefully evolved network. Neural Computation 1:253–262.

    Google Scholar 

  • Caudill, M.1988. Neural networks primer PartV. AI Expert, Nov.:57-65.

  • Connell, J. and Mahadevan, S. 1993. Robot Learning. Kluwer Academic Publishers.

  • Dorigo, M. (Ed.). 1996. Special issue on learning autonomous robots. IEEE Transactions on Systems Man and Cybernetics, Part B: Cybernetics23:(3).

  • Drickamer, L.C., Vessey, S.H., and Meikle, D. 1996. Animal Behaviour: Mechanisms, Ecology, Evolution. Wm. C. Brown Publishers.

  • Franklin, J., Mitchell, T., and Thrun, S. (Eds.). 1996. Robot Learning-Special Issue of Machine Learning Journal. Kluwer Academic Press, vol. 23, no. 2/3.

  • Gat, E., Desai, R., Ivlev, R., Loch, J., and Miller, D.P. 1994. Behavior control for robotic exploration of planetary surfaces. IEEE Transactions on Robotics and Automation, 10(4): 490–503.

    Google Scholar 

  • Gordon, M.S. 1994. Animal Physiology: Principles and Adaptations, 2nd edn., Macmillan Publishing Co. Inc.: New York, 1972.

    Google Scholar 

  • Held, R. and Hein, A. 1963. Movement produced stimulation in the development of visually guided behavior, Journal of Comparative Physiology and Psychology, 56:872–876.

    Google Scholar 

  • Kaelbling, L.P. 1993. Learning in Embedded Systems. MIT Press.

  • Mahadevan, S. and Theocharus, G. 1998. Rapid concept learning for mobile robots, Autonomous Robots, 5:(3/4), pp. 239–251 Kluwer Academic Publishers.

    Google Scholar 

  • Michaud, F. and Mataric, M.J. 1998. Learning from history for behavior-based mobile robots in non-stationary conditions. Autonomous Robots, 5(3/4):335–354.

    Google Scholar 

  • Minsky, M. 1963. Steps towards artificial intelligence. In Computers and Thought, Edward A Feigenbaum & Julian Feldman, McGraw-Hill: N.Y.

    Google Scholar 

  • Pfeifer, R. 1996. Building “fungus eaters”; design principles of autonomous agents. In Proceedings of the Fourth International Conference on Simulation of Adaptive Behavior. The MIT Press: Cambridge, Mass, pp. 3–12.

    Google Scholar 

  • Porr, B. and Wörgötter, F. 2001. Temporal Hebbian learning in rate-coded neural networks: a theoretical approach towards classical conditioning. In Proceedings of Artifical Neural Networks-ICANN 2001, Springer, August 21-25, pp. 1115-1120.

  • Scott, J.P. 1972. Animal Behavior, 2nd edn. The University of Chicago Press: Chicago.

    Google Scholar 

  • Sharkey, N.E. 1998. Learning from innate behaviors: A quantitative evaluation of neural network controllers, Autonomous Robots, 5 (3/4):317–334.

    Google Scholar 

  • Steels, L. 1994. Case study in the behavior-oriented design of autonomous agents, From Animals to Animats 3. In Proc. 3rd International Conference on the Simulation of Adaptive Behavior, MIT Press, pp. 445-452.

  • Smart, W.D. and Kaelbling, L.P., 2002. Effective reinforcement learning for mobile robots, In Proceedings of the IEEE International Conference on Robotics and Automation Washington D.C., pp. 3404-3410.

  • Turing A.M. 1950. Computing machines and intelligence. Mind, 49:(236):433–460.

    Google Scholar 

  • Walter, W.G. 1950. An Imitation of Life. Scientific American, May:42-45.

  • Walter, W.G. 1963. The Living Brain. Penguin Books Ltd.

Download references

Authors

Rights and permissions

Reprints and permissions

About this article

Cite this article

Russell, R.A. Mobile Robot Learning by Self-Observation. Autonomous Robots 16, 81–93 (2004). https://doi.org/10.1023/B:AURO.0000008672.32974.bb

Download citation

  • Issue Date:

  • DOI: https://doi.org/10.1023/B:AURO.0000008672.32974.bb

Navigation