Skip to main content

Humanoid Active Audition System Improved by the Cover Acoustics

  • Conference paper
PRICAI 2000 Topics in Artificial Intelligence (PRICAI 2000)

Part of the book series: Lecture Notes in Computer Science ((LNAI,volume 1886))

Included in the following conference series:

Abstract

Perception system for humanoid should be active, e.g., by moving its body or controlling parameters of sensors such as cameras or microphones, to perceive environments better. This paper focuses on active audition, whose main problem is to suppress internal sounds made by humanoid movements. Otherwise, such sounds would deteriorate the performance of auditory processing. Our 4-degree-of-freedom (DOF) humanoid, called SIG, has a cover to enclose internal sounds from the outside. SIG has a pair of left and right microphones to collect internal sounds and another pair to collect external sounds originating from the outside of SIG. A simple strategy of choosing a subband of external sounds if sounds from internal microphones in the same subband is weaker than those from external microphones sometimes fails in internal sound cancellation due to resonance within the cover. In this paper, we report the design of internal sound cancellation system to enhance external sounds. First, the acoustic characteristic of the humanoid cover is measured to make a model of each motor movement. Then, an adaptive filter is designed based on the model by taking movement commands into accounts. Experiments show that this cancellation system enhances external sounds and SIG can track and localize sound sources during its movement.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 84.99
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 109.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. Y. Aloimonos, I. Weiss, and A. Bandyopadhyay. Active vision. International Journal of Computer Vision, 1(4):333–356, 1987.

    Article  Google Scholar 

  2. S. F. Boll. A spectral subtraction algorithm for suppression of acoustic noise in speech. In Proceedings of 1979 International Conference on Acoustics, Speech, and Signal Processing (ICASSP-79), pages 200–203. IEEE, 1979.

    Google Scholar 

  3. R. Brooks, C. Breazeal, M. Marjanovie, B. Scassellati, and M. Williamson. The cog project: Building a humanoid robot. Technical report, MIT, 1999.

    Google Scholar 

  4. G. J. Brown. Computational auditory scene analysis: A representational approach. PhD thesis, Dept. of Computer Science, University of Sheffield, 1992.

    Google Scholar 

  5. J. Cavaco, S. ad Hallam. A biologically plausible acoustic azimuth estimation system. In Proceedings of IJCAI-99 Workshop on Computational Auditory Scene Analysis (CASA’99), pages 78–87. IJCAI, Aug. 1999.

    Google Scholar 

  6. M. P. Cooke, G. J. Brown, M. Crawford, and P. Green. Computational auditory scene analysis: Listening to several things at once. Endeavour, 17(4):186–190, 1993.

    Article  Google Scholar 

  7. O. D. Faugeras. Three Dimensional Computer Vision: A Geometric Viewpoint. The MIT Press, MA., 1993.

    Google Scholar 

  8. M. Kawato. Bi-directional theory approach to consciousness. In Cognition, Computation, and Consciousness. Oxford University Press, 1996.

    Google Scholar 

  9. N. Kita, S. Rougeaux, Y. Kuniyoshi, and S. Sakane. Real-time binocular tracking based on virtual horopter. Journal of Robotics Society Japan, 13(5):101–108, 1995.

    Google Scholar 

  10. H. Kitano, H. G. Okuno, K. Nakadai, I. Fermin, T. Sabish, Y. Nakagawa, and T. Matsui. Designing a humanoid head for robocup challenge. In Proceedings of 4th International Conference on Autonomous Agents (Agents 2000). ACM, 2000.

    Google Scholar 

  11. Y. Matsusaka, T. Tojo, S. Kuota, K. Furukawa, D. Tamiya, K. Hayata, Y. Nakano, and T. Kobayashi. Multi-person conversation via multi-modal interface-a robot who communicates with multi-user. In Proceedings of Eurospeech, pages 1723–1726. ESCA, 1999.

    Google Scholar 

  12. K. Nakadai, T. Lourens, H. G. Okuno, and H. Kitano. Active audition for humanoid. In Proceedings of 17th National Conference on Artificial Intelligence (AAAI-2000). AAAI, 2000. (to appear).

    Google Scholar 

  13. K. Nakadai, T. Matsui, H. G. Okuno, and H. Kitano. Active audition system and humanoid exterior design. In Proceedings of International Conference on Intelligent Robots and Systems (IROS 2000). IEEE, 2000. (accepted).

    Google Scholar 

  14. K. Nakadai, H. G. Okuno, and H. Kitano. A method of peak extraction and its evaluation for humanoid. In SIG-Challenge-99-7, pages 53–60. JSAI, 1999.

    Google Scholar 

  15. Y. Nakagawa, H. G. Okuno, and H. Kitano. Using vision to improve sound source separation. In Proceedings of 16th National Conference on Artificial Intelligence (AAAI-99), pages 768–775. AAAI, 1999.

    Google Scholar 

  16. T. Nakatani, H. G. Okuno, and T. Kawabata. Auditory stream segregation in auditory scene analysis with a multi-agent system. In Proceedings of 12th National Conference on Artificial Intelligence (AAAI-94), pages 100–107. AAAI, 1994.

    Google Scholar 

  17. T. Nakatani, H. G. Okuno, and T. Kawabata. Residue-driven architecture for computational auditory scene analysis. In Proceedings of 14th International Joint Conference on Artificial Intelligence (IJCAI-95), volume 1, pages 165–172. AAAI, 1995.

    Google Scholar 

  18. D. Rosenthal and H. G. Okuno, editors. Computational Auditory Scene Analysis. Lawrence Erlbaum Associates, NJ., 1998.

    Google Scholar 

  19. A. Takanishi, S. Masukawa, Y. Mori, and T. Ogawa. Development of an anthropomorphic auditory robot that localizes a sound direction (in Japanese). Bulletin of the Centre for Informatics, 20:24–32, 1995.

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2000 Springer-Verlag Berlin Heidelberg

About this paper

Cite this paper

Nakadai, K., Okuno, H.G., Kitano, H. (2000). Humanoid Active Audition System Improved by the Cover Acoustics. In: Mizoguchi, R., Slaney, J. (eds) PRICAI 2000 Topics in Artificial Intelligence. PRICAI 2000. Lecture Notes in Computer Science(), vol 1886. Springer, Berlin, Heidelberg. https://doi.org/10.1007/3-540-44533-1_55

Download citation

  • DOI: https://doi.org/10.1007/3-540-44533-1_55

  • Publisher Name: Springer, Berlin, Heidelberg

  • Print ISBN: 978-3-540-67925-7

  • Online ISBN: 978-3-540-44533-3

  • eBook Packages: Springer Book Archive

Publish with us

Policies and ethics