Skip to main content
Log in

Construction of a robot control system which understands a voice and pointing actions

  • Original Article
  • Published:
Artificial Life and Robotics Aims and scope Submit manuscript

Abstract

This research was aimed at making a robot that can go and get an object designated by the user. We produced a robot control system that uses pointing and voice control. This control system is composed of two systems. One system is the object instruction system that uses pointing, and the other is the object instruction system that uses voice. The approximate position of a designated object is recognized by the object instruction system that uses pointing, and information about the object and instructions about the operation are conveyed by the object instruction system that uses voice. This time, the object instruction system that uses pointing has been designed and verified as the first step. A calculation resource to calculate the approximate position of the designated object is obtained when the user points directly at the object. A robot can be moved around a designated object by using this system. The object instruction system that uses voice will be constructed in future work.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Similar content being viewed by others

References

  1. Inaba M, Kagami S, Kanehiro F, et al (2000) Remote-brained robot approach. Int J Robotics Res 19:933–954

    Article  Google Scholar 

  2. Watanabe Y, Abe N, Tanaka K, et al (2005) Multimodal communication system allowing man and avatar to use voice and beck. 3rd International Conference on Information Technology and Applications (ICITA’2005), pp 161–166

  3. Goto H, Abe N, Tanaka K, et al (2005) Usage of shared memory for multimodal data processing in an outdoor mobile robot. Proceedings of the ISR (36th International Symposium on Robotics) (CD-ROM)

  4. Bellamine M, Abe N, Tanaka K, et al (2005) Remote diagnosis system for rotating machinery using virtual reality. IEICE Trans, pp 895–903

  5. Oomori T, Abe N, Tanaka K, et al (2005) Concurrent development of virtual robots and real robots based on physical law. IMEKO/IFAC/IFIP Workshop on Advanced Robot Systems and Virtual Reality (CD-ROM)

  6. Tabuchi Y, Abe N, Taki H, et al (2008) Synchronization between audiovisual and haptic feeling for constructing edutainment systems. Virtual Reality 12:27–36

    Article  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Yuki Takenaka.

Additional information

This work was presented in part at the 14th International Symposium on Artificial Life and Robotics, Oita, Japan, February 5–7, 2009

About this article

Cite this article

Takenaka, Y., Abe, N., Tabuchi, Y. et al. Construction of a robot control system which understands a voice and pointing actions. Artif Life Robotics 14, 224 (2009). https://doi.org/10.1007/s10015-009-0658-1

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1007/s10015-009-0658-1

Key words

Navigation