Abstract
In this paper, we propose a method to apply robotic imitation learning in robot teams. In our method, behavior primitives with task-relevant information are defined as the basic units for robots to complete a task. Each behavior has its own task-relevant affordances. The learned behavior primitives format TAEM based behavior libraries and are stored in a database for robots to share. The motivation of learning, conducted by robots, is goal-oriented and strongly related to the given task. Given a task, a robot analyzes the environment and searches the behavior library to find suitable behaviors to generate a behavior sequence to complete the task. If it thinks that it cannot complete this task, this robot requests other robots for assistance or request a human teacher to demonstrate the required behaviors. The newly learned behaviors will be added into the existing behavior library. We also develop inhibiting properties for robots to evaluate the current behaviors, which enables robots to request collaborations from other robots. The experimental results show the validity our proposed method.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
References
Atkeson, C., Schaal, S.: Robot Learning from Demonstration. In: Fisher Jr., D.H. (ed.) The Fourteenth International Conference on Machine Learning, pp. 11–73. Morgan Kaufmann (1997)
Billard, A., Calinon, S., Dillmann, R., Schaal, S.: Robot programming by demonstration. In: Siciliano, B., Khatib, O. (eds.) Handbook of Robotics. Springer, New York (2007)
Bitzer, S., Vijayakumar, S.: Latent Spaces for Dynamic Movement Primitives. In: The 2009 IEEE-RAS International Conference on Humanoid Robots, pp. 574–581 (2009)
Calinon, S., Guenter, F., Billard, A.: On learning, representing, and generalizing a task in a humanoid robot. IEEE Transactions on Systems, Man, and Cybernetics, Part B 37, 286–298 (2007)
Cormen, T.H., Leiserson, C.E., Rivest, R.L., Stein, C.: Introduction to algorithms. MIT Press (2001)
Dillmann, R., Rogalla, O., Ehrenmann, M., Zollner, R., Bordegoni, M.: Learning robot behaviour and skills based on human demonstration and advice: the machine learning paradigm. In: Ninth International Symposium of Robotics Research (ISRR 1999), Citeseer, Snowbird, UT, USA, vol. 9, pp. 229–238 (2000)
Ijspeert, A., Nakanishi, J., Schaal, S.: Learning attractor landscapes for learning motor primitives. In: Becker, S., Thrun, S., Obermayer, K. (eds.) Advances in Neural Information Processing Systems, vol. 15, pp. 1547–1554. MIT Press (2003)
Ijspeert, A., Nakanishi, J., Schaal, S.: Movement imitation with nonlinear dynamical systems in humanoid robots. In: 2002 IEEE International Conference on Robotics and Automation, Citeseer, Washington, DC, USA, vol. 2, pp. 1398–1403 (2002)
Lesser, V., Decker, K., Wagner, T., Carver, N., Garvey, A., Horling, B., et al.: Evolution of the GPGP/TAEMS domain-independent coordination framework. Autonomous Agents and Multi-agent Systems 9, 87–143 (2004)
Rasmussen, C.: Gaussian processes in machine learning. In: Advanced Lectures on Machine Learning, pp. 63–71. MIT Press, Cambridge (2004)
Tan, H.: Implementation of a Framework for Imitation Learning on a Humanoid Robot using a Cognitive Architecture. In: Zaier, R. (ed.) The Future of Humanoid Robots: Research and Applications, pp. 189–210. InTech (2012)
Tan, H., Du, Q., Wu, N.: A Framework for Cognitive Robots to Learn Behaviors through Imitation and Interaction with Humans. In: 2012 IEEE International Multi-Disciplinary Conference on Cognitive Methods in Situation Awareness and Decision Support, New Orleans, USA, pp. 235–238 (2012a)
Tan, H., Du, Q., Wu, N.: Robots Learn Writing. Journal of Robotics (2012b)
Tan, H., Erdemir, E., Kawamura, K., Du, Q.: A Potential Field Method-based Extension of the Dynamic Movement Primitive Algorithm for Imitation Learning with Obstacle Avoidance. In: 2011 IEEE International Conference on Mechatronics and Automation, Beijing, China, pp. 525–530 (2011)
Tan, H., Kawamura, K.: A Computational Framework for Integrating Robotic Exploration and Human Demonstration in Imitation Learning. In: 2011 IEEE International Conference on System, Man and Cybernetics, Anchorage, AK, USA, pp. 2501–2506 (2011)
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2014 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Tan, H. (2014). Imitation Learning and Behavior Generation in a Robot Team. In: Ani Hsieh, M., Chirikjian, G. (eds) Distributed Autonomous Robotic Systems. Springer Tracts in Advanced Robotics, vol 104. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-55146-8_30
Download citation
DOI: https://doi.org/10.1007/978-3-642-55146-8_30
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-642-55145-1
Online ISBN: 978-3-642-55146-8
eBook Packages: EngineeringEngineering (R0)