Abstract
This research was aimed at making a robot that can go and get an object designated by the user. We produced a robot control system that uses pointing and voice control. This control system is composed of two systems. One system is the object instruction system that uses pointing, and the other is the object instruction system that uses voice. The approximate position of a designated object is recognized by the object instruction system that uses pointing, and information about the object and instructions about the operation are conveyed by the object instruction system that uses voice. This time, the object instruction system that uses pointing has been designed and verified as the first step. A calculation resource to calculate the approximate position of the designated object is obtained when the user points directly at the object. A robot can be moved around a designated object by using this system. The object instruction system that uses voice will be constructed in future work.
Similar content being viewed by others
Explore related subjects
Discover the latest articles, news and stories from top researchers in related subjects.References
Inaba M, Kagami S, Kanehiro F, et al (2000) Remote-brained robot approach. Int J Robotics Res 19:933–954
Watanabe Y, Abe N, Tanaka K, et al (2005) Multimodal communication system allowing man and avatar to use voice and beck. 3rd International Conference on Information Technology and Applications (ICITA’2005), pp 161–166
Goto H, Abe N, Tanaka K, et al (2005) Usage of shared memory for multimodal data processing in an outdoor mobile robot. Proceedings of the ISR (36th International Symposium on Robotics) (CD-ROM)
Bellamine M, Abe N, Tanaka K, et al (2005) Remote diagnosis system for rotating machinery using virtual reality. IEICE Trans, pp 895–903
Oomori T, Abe N, Tanaka K, et al (2005) Concurrent development of virtual robots and real robots based on physical law. IMEKO/IFAC/IFIP Workshop on Advanced Robot Systems and Virtual Reality (CD-ROM)
Tabuchi Y, Abe N, Taki H, et al (2008) Synchronization between audiovisual and haptic feeling for constructing edutainment systems. Virtual Reality 12:27–36
Author information
Authors and Affiliations
Corresponding author
Additional information
This work was presented in part at the 14th International Symposium on Artificial Life and Robotics, Oita, Japan, February 5–7, 2009
About this article
Cite this article
Takenaka, Y., Abe, N., Tabuchi, Y. et al. Construction of a robot control system which understands a voice and pointing actions. Artif Life Robotics 14, 224 (2009). https://doi.org/10.1007/s10015-009-0658-1
Received:
Accepted:
Published:
DOI: https://doi.org/10.1007/s10015-009-0658-1