Abstract
This paper investigates the credibility of voice (especially natural language commands) as a communication medium in sharing advanced sensory capacity and knowledge of the human with a robot to perform a cooperative task. Identification of the machine sensitive words in the unconstrained speech signal and interpretation of the imprecise natural language commands for the machine has been considered. The system constituents include a hidden Markov model (HMM) based continuous automatic speech recognizer (ASR) to identify the lexical content of the user’s speech signal, a fuzzy neural network (FNN) to comprehend the natural language (NL) contained in identified lexical content, an artificial neural network (ANN) to activate the desired functional ability, and control modules to generate output signals to the actuators of the machine. The characteristic features have been tested experimentally by utilizing them to navigate a Khepera® in real time using the user’s visual information transferred by speech signals.
Preview
Unable to display preview. Download preview PDF.
Similar content being viewed by others
References
Mazo M., Rodrìguez F. J. et al. (1995) Electronic control of a wheelchair guided by voice commands. Control Eng. Practice 3(5):665–674
Sugisaka M., Fan X. (2001) Control of a welfare life robot guided by voice commands. In: Proc. of the ICCAS 2001, Cheju Korea, 390–393
Lin C. T., Kan M. C. (1998) Adaptive fuzzy command acquisition with reinforcement learning. IEEE Transaction on Fuzzy Systems 6(1): 102–121
Komiya K., Morita K. et al. (2000) Guidence of a wheelchair by voice. In: IECON 2000, Nagoya Japan, 102–107
Pulasinghe K., Watanabe K. et al. (2001) Modular fuzzy neural controller driven by voice commands. In: Proc. of the ICCAS 2001, Cheju Korea, 194–197
Rose R. C, Paul D. B. (1990) A hidden Markov model based keyword recognition system. In: Proc. of the IEEE ICASSP ’90, Albuquerque New Mexico, 129–132
Jeanrenaud P., Ng K. et al. (1993) Phonetic-based word spotter: Various configurations and application to event spotting. In: Proc. of the EUROSPEECH ’93, Berlin Germany, 1057–1060
Bazzi I., Glass J. R. (2000) Modeling out-of-vocabulary words for robust speech recognition. In: Proc. of the ICSLP 2000, Beijing China.
Leeuwen D. A. V., Kraaij W. et al. (1999) Prediction of keywords spotting performance based on phonemic contents. In: Proc. of the ESCA/ETRW, Cambridge UK, 73–77
Young S. J. (1993) The HTK hidden Markov model toolkit: Design and philosophy. Technical Report TR.153, Department of Engineering, Cambridge University UK
Lippmann R. P. (1987) An introduction to computing with neural nets. IEEE Magazine on Acoustics, Signal, and Speech Processing 4:4–22
Jang J. S. R., Sun C. T. (1995) Neuro-fuzzy modeling and control. In: Proc. of the IEEE 83(3):378–406
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2002 Springer-Verlag Tokyo
About this paper
Cite this paper
Pulasinghe, K., Watanabe, K., Kiguchi, K., Izumi, K. (2002). Voice Communication in Performing a Cooperative Task with a Robot. In: Asama, H., Arai, T., Fukuda, T., Hasegawa, T. (eds) Distributed Autonomous Robotic Systems 5. Springer, Tokyo. https://doi.org/10.1007/978-4-431-65941-9_13
Download citation
DOI: https://doi.org/10.1007/978-4-431-65941-9_13
Publisher Name: Springer, Tokyo
Print ISBN: 978-4-431-65943-3
Online ISBN: 978-4-431-65941-9
eBook Packages: Springer Book Archive