Abstract
Communication between humans and robots is crucial to achieve successful cooperation in real-life scenarios. The robot must understand not only linguistic expressions, but also non-linguistic expressions such as nodding and gestures. In this research, we examine whether a listener nods in response to a speaker’s utterance. Our proposed method judges nodding based on the movement of the listener’s facial keypoints and the speaker’s speech intonation. The proposed method achieves approximately 84.4% recognition accuracy when we input the movement and intonation simultaneously. This improves nodding recognition accuracy by 8.8% over movement only approach. This result indicates that the movement of the listener’s facial keypoints and the speaker’s intonation are important information in nodding recognition.
This is a preview of subscription content, log in via an institution.
Buying options
Tax calculation will be finalised at checkout
Purchases are for personal use only
Learn about institutional subscriptionsReferences
Kobayashi, N., et al.: Quantitative evaluation of infant behavior and mother infant interaction. Early Dev. Parent. 1(1), 23–31 (1992)
Graves, A., et al.: Speech recognition with deep recurrent neural networks. In: Acoustics, Speech and Signal Processing (ICASSP), pp. 6645–6649 (2013)
Hochreiter, S., et al.: Long short-term memory. Neural Comput. 9(8), 1735–1780 (1997)
Simonyan, K., et al.: Two-stream convolutional networks for action recognition in videos. NIPS (2014)
Wu, L., et al.: In Vivo evaluation of wearable head impact sensors. Ann. Biomed. Eng. 44(4), 1234–45 (2015)
King, D.E.: Dlib-ml: a machine learning toolkit. J. Mach. Learn. Res. 10, 1755–1758 (2009)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2019 Springer Nature Switzerland AG
About this paper
Cite this paper
Yamashita, T., Nakagawa, M., Fujiyoshi, H., Haikawa, Y. (2019). Recognition of Listener’s Nodding by LSTM Based on Movement of Facial Keypoints and Speech Intonation. In: Stephanidis, C. (eds) HCI International 2019 - Posters. HCII 2019. Communications in Computer and Information Science, vol 1033. Springer, Cham. https://doi.org/10.1007/978-3-030-23528-4_22
Download citation
DOI: https://doi.org/10.1007/978-3-030-23528-4_22
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-23527-7
Online ISBN: 978-3-030-23528-4
eBook Packages: Computer ScienceComputer Science (R0)