Abstract
Perception system for humanoid should be active, e.g., by moving its body or controlling parameters of sensors such as cameras or microphones, to perceive environments better. This paper focuses on active audition, whose main problem is to suppress internal sounds made by humanoid movements. Otherwise, such sounds would deteriorate the performance of auditory processing. Our 4-degree-of-freedom (DOF) humanoid, called SIG, has a cover to enclose internal sounds from the outside. SIG has a pair of left and right microphones to collect internal sounds and another pair to collect external sounds originating from the outside of SIG. A simple strategy of choosing a subband of external sounds if sounds from internal microphones in the same subband is weaker than those from external microphones sometimes fails in internal sound cancellation due to resonance within the cover. In this paper, we report the design of internal sound cancellation system to enhance external sounds. First, the acoustic characteristic of the humanoid cover is measured to make a model of each motor movement. Then, an adaptive filter is designed based on the model by taking movement commands into accounts. Experiments show that this cancellation system enhances external sounds and SIG can track and localize sound sources during its movement.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
References
Y. Aloimonos, I. Weiss, and A. Bandyopadhyay. Active vision. International Journal of Computer Vision, 1(4):333–356, 1987.
S. F. Boll. A spectral subtraction algorithm for suppression of acoustic noise in speech. In Proceedings of 1979 International Conference on Acoustics, Speech, and Signal Processing (ICASSP-79), pages 200–203. IEEE, 1979.
R. Brooks, C. Breazeal, M. Marjanovie, B. Scassellati, and M. Williamson. The cog project: Building a humanoid robot. Technical report, MIT, 1999.
G. J. Brown. Computational auditory scene analysis: A representational approach. PhD thesis, Dept. of Computer Science, University of Sheffield, 1992.
J. Cavaco, S. ad Hallam. A biologically plausible acoustic azimuth estimation system. In Proceedings of IJCAI-99 Workshop on Computational Auditory Scene Analysis (CASA’99), pages 78–87. IJCAI, Aug. 1999.
M. P. Cooke, G. J. Brown, M. Crawford, and P. Green. Computational auditory scene analysis: Listening to several things at once. Endeavour, 17(4):186–190, 1993.
O. D. Faugeras. Three Dimensional Computer Vision: A Geometric Viewpoint. The MIT Press, MA., 1993.
M. Kawato. Bi-directional theory approach to consciousness. In Cognition, Computation, and Consciousness. Oxford University Press, 1996.
N. Kita, S. Rougeaux, Y. Kuniyoshi, and S. Sakane. Real-time binocular tracking based on virtual horopter. Journal of Robotics Society Japan, 13(5):101–108, 1995.
H. Kitano, H. G. Okuno, K. Nakadai, I. Fermin, T. Sabish, Y. Nakagawa, and T. Matsui. Designing a humanoid head for robocup challenge. In Proceedings of 4th International Conference on Autonomous Agents (Agents 2000). ACM, 2000.
Y. Matsusaka, T. Tojo, S. Kuota, K. Furukawa, D. Tamiya, K. Hayata, Y. Nakano, and T. Kobayashi. Multi-person conversation via multi-modal interface-a robot who communicates with multi-user. In Proceedings of Eurospeech, pages 1723–1726. ESCA, 1999.
K. Nakadai, T. Lourens, H. G. Okuno, and H. Kitano. Active audition for humanoid. In Proceedings of 17th National Conference on Artificial Intelligence (AAAI-2000). AAAI, 2000. (to appear).
K. Nakadai, T. Matsui, H. G. Okuno, and H. Kitano. Active audition system and humanoid exterior design. In Proceedings of International Conference on Intelligent Robots and Systems (IROS 2000). IEEE, 2000. (accepted).
K. Nakadai, H. G. Okuno, and H. Kitano. A method of peak extraction and its evaluation for humanoid. In SIG-Challenge-99-7, pages 53–60. JSAI, 1999.
Y. Nakagawa, H. G. Okuno, and H. Kitano. Using vision to improve sound source separation. In Proceedings of 16th National Conference on Artificial Intelligence (AAAI-99), pages 768–775. AAAI, 1999.
T. Nakatani, H. G. Okuno, and T. Kawabata. Auditory stream segregation in auditory scene analysis with a multi-agent system. In Proceedings of 12th National Conference on Artificial Intelligence (AAAI-94), pages 100–107. AAAI, 1994.
T. Nakatani, H. G. Okuno, and T. Kawabata. Residue-driven architecture for computational auditory scene analysis. In Proceedings of 14th International Joint Conference on Artificial Intelligence (IJCAI-95), volume 1, pages 165–172. AAAI, 1995.
D. Rosenthal and H. G. Okuno, editors. Computational Auditory Scene Analysis. Lawrence Erlbaum Associates, NJ., 1998.
A. Takanishi, S. Masukawa, Y. Mori, and T. Ogawa. Development of an anthropomorphic auditory robot that localizes a sound direction (in Japanese). Bulletin of the Centre for Informatics, 20:24–32, 1995.
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2000 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Nakadai, K., Okuno, H.G., Kitano, H. (2000). Humanoid Active Audition System Improved by the Cover Acoustics. In: Mizoguchi, R., Slaney, J. (eds) PRICAI 2000 Topics in Artificial Intelligence. PRICAI 2000. Lecture Notes in Computer Science(), vol 1886. Springer, Berlin, Heidelberg. https://doi.org/10.1007/3-540-44533-1_55
Download citation
DOI: https://doi.org/10.1007/3-540-44533-1_55
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-540-67925-7
Online ISBN: 978-3-540-44533-3
eBook Packages: Springer Book Archive