Skip to main content
Log in

Interactive learning and management of visual information via human-like software robot

  • Special Issue
  • Published:
New Generation Computing Aims and scope Submit manuscript

Abstract

To achieve smooth real-world interaction between people and computers, we developed a system that displays a three-dimensional computer-graphic human-like image from the waist up (anthropomorphic software robot: hereinafter “robot”) on the display, that interactively sees and hears, and that has fine and detailed control functions such as facial expressions, line of sight, and pointing at targets with its finger. The robot visually searches and identifies persons and objects in real space that it has learned in advance (registered space, which was our office in this case), manages the history information of the places and times it found objects and/or persons, and tells the user, indicating their three-dimensional positions with line of sight and its finger. It interactively learns new objects and persons with line of with their names and owners. By using this function, the robot can engage in simple dialogue (do a task) with the user.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Similar content being viewed by others

References

  1. Akaho S., Hayamizu, S., Hasegawa O., Yoshimura T. and Asoh H., “Concept Acquiaition from Multiple Information Source by the EM Algorithm,”Trans. of the IEICE, A, J80-A, 9, pp. 1546–1553, 1997. (In Japanese)

    Google Scholar 

  2. Brooks, R.A., “A Robust Layered Control System for a Mobile Robot,”IEEE Journal of Robotics and Automation, March, 1986.

  3. (For example) Brooks R.A. and Stein L.A., “Building Brains for Bodies,”Autonomous Robots, 1, 1, pp. 7–25, 1994.

    Article  Google Scholar 

  4. ETL CG Tool: http://www.etl.go.jp/etl/gazo/CGtool/

  5. Dreyfus H.,What Computers Still Can’t Do MIT Press, Cambridge, 1993.

    Google Scholar 

  6. Hasegawa O., Morishima S. and Kaneko M., “Processing of Facial Information by Computer,”ELECTRONICS and COMMUNICATIOS in Japan, 3, 81, 10, pp. 36–57, Scripta Technica, Inc., 1998.

    Article  Google Scholar 

  7. Hasegawa, O., Yokosawa K., Ishizuka M., “Realtime Face Image Recognition in Unconstrained Environment for Interactive Visual Interface,”Proc. 12th ICPR (Int’l Conference on Pattern Recognition), pp. 384–387, 1994.

  8. Hasegawa O., “Computer Vision Inspired by Biological Visual Systems,”Journal of the IPSJ (The Information Processing Society of Japan), 39, 2, pp. 133–138, 1998. (In Japanese)

    Google Scholar 

  9. Hasegawa O., Itou K., Kurita T., Hayamizu S., Tanaka K., Yamamoto K. and Otsu N., “Active Agent Oriented Multimodal Interface System,”Proc. IJCAI-95 (14th Int’l Joint Conf. on Artificial Intelligence), pp. 82–87, 1995.

  10. Itou K., Hayamizu S. and Tanaka K., “System design, data collection and evaluation of a speech dialogue system,”IEICE Trans, INF. & SYST. E76-D., 1, pp. 121–127, 1993.

    Google Scholar 

  11. Otsu N. and Kurita T., “A new scheme for practical flexible and intelligent vision systems,”Proc. IAPR Workshop on Computer Vision, Tokyo, pp. 431–435, 1988.

  12. Raytchev B., Hasegawa O. and Otsu N., “Online Gesture Recogniton Using Predictive Statistical Feature Extraction and Multivariate Analysis,”Proc IAPR Workshop on Machine Vision Applications (MVA), pp. 569–572, 1998.

  13. Russell S. and Norving P.,Artificial Intelligence: A Modern Approach, Prentice Hall, 1995.

  14. Shimojo S.,Beginning to See — Genesis of Mind, and The New Infant Science, Shinyosya, Tokyo, 1988. (In Japanese)

    Google Scholar 

  15. Singh J. A. L. and Zingg R. M.,Wolf-Children and Feral Man, Harper & Brothers, 1942.

  16. Taylor C. E., Langton C. G. and Kitano H. ed,Artificial Life, MIT Press, Cambridge

  17. Yoda I. and Sakaue K., “Utilization of Stereo Disparity and Optical Flow Information for Human Interaction,”Proc. 6th International Conference on Computer Vision, pp. 1109–1114, 1998.

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Osamu Hasegawa.

Additional information

Osamu Hasegawa, Ph.D.: He received the B.E. and M.E. degrees from the Science University of Tokyo, in 1988, 1990 respectively. He received Ph.D. degree from the University of Tokyo, in 1993. Currently, he is a senior research scientist at the Electrotechnical Laboratory (ETL), Tsukuba, Japan. His research interests include Computer Vision and Multi-modal Human Interface. Dr. Hasegawa is a member of the AAAI, the Institute of Electronics, Information and Communication Engineers, Japan (IEICE), Information Processing Society of Japan and others.

Katsuhiko Sakaue, Ph.D.: He received the B.E., M.E., and Ph.D. degrees all in electronic engineering from the University of Tokyo, in 1976, 1978 and 1981, respectively. In 1981, he joined the Electrotechnical Laboratory, Ministry of International Trade and Industry, and engaged in researches in image processing and computer vision. He received the Encouragement Prize in 1979 from IEICE, and the Paper Award in 1985 from Information Processing Society of Japan. He is a member of IEICE, IEEE, IPSJ, ITE.

Satoru Hayamizu, Ph.D.: He is a leader of Interactive Intermodal Integration Lab. at Electrotechnical Laboratory. He received the B.E., M.E., Ph.D. degrees from Tokyo University. Since 1981, he has been working on speech recognition, spoken dialogue, and communication with artifacts. From 1989 to 1990, he was a visiting scholar in Carnegie Mellon University and in 1994 a visiting scientist in LIMSI/CNRS.

About this article

Cite this article

Hasegawa, O., Sakaue, K. & Hayamizu, S. Interactive learning and management of visual information via human-like software robot. New Gener Comput 18, 103–116 (2000). https://doi.org/10.1007/BF03037589

Download citation

  • Received:

  • Revised:

  • Issue Date:

  • DOI: https://doi.org/10.1007/BF03037589

Keywords

Navigation