ABSTRACT
In this demonstration, we will showcase realtime grounded language learning on the humanoid robot Pepper. In particular, learning word-object and word-action mapping from cross-modal data, where simple actions, such as take, put and push, are shown to the robot by a human tutor. The visual demonstration is accompanied by verbal descriptions of the performed actions, such as I take the box and put it next to the bottle. Learning was realized on the humanoid robot Pepper using the Google Speech API for speech to text and the robot's camera system for object tracking.
- Amir Aly and Tadahiro Taniguchi. 2018. Towards Understanding Object-Directed Actions: A Generative Model for Grounding Syntactic Categories of Speech through Visual Perception. In Proc. of the IEEE International Conference on Robotics and Automation (ICRA) 2018. IEEE.Google ScholarCross Ref
- Johann Prankl, Aitor Aldoma, Alexander Svejda, and Markus Vincze. 2015. RGB-D object modelling for object recognition and tracking. In Intelligent Robots and Systems (IROS), 2015 IEEE/RSJ International Conference on. IEEE, 96--103.Google ScholarCross Ref
- Jun Tani, Masato Ito, and Yuuya Sugita. 2004. Self-organization of distributedly represented multiple behavior schemata in a mirror system: reviews of robot experiments using RNNPB. Neural Networks 17, 8-9 (2004), 1273--1289. Google ScholarDigital Library
Index Terms
- Grounded Word Learning on a Pepper Robot
Recommendations
Acquiring Mobile Robot Behaviors by Learning Trajectory Velocities
The development of robots that learn from experience is a relentless challenge confronting artificial intelligence today. This paper describes a robot learning method which enables a mobile robot to simultaneously acquire the ability to avoid objects, ...
Autonomous Depth Perception of Humanoid Robot Using Binocular Vision System Through Sensorimotor Interaction with Environment
ICONIP 2015: Proceeings, Part II, of the 22nd International Conference on Neural Information Processing - Volume 9490In this paper, we explore how a humanoid robot having two cameras can learn to improve depth perception by itself. We propose an approach that can autonomously improve depth estimation of the humanoid robot. This approach can tune parameters that are ...
Bimanual Haptics for Humanoid Robot Teleoperation Using ROS and V-REP
ICARSC '15: Proceedings of the 2015 IEEE International Conference on Autonomous Robot Systems and CompetitionsThe high degree of complexity associated with humanoid robotic platforms significantly hinders the definition of accurate models. Robot learning based on human teleoperation provides suitable ways and means of meeting this challenge. Teleoperating the ...
Comments