ABSTRACT
It is necessary for humanoid robots to have the capability of human-like social environment perception, which could enable robots to interact with people intelligently. In order to improve the interaction capability of humanoid robots in complex and changeable social environments, an interactive intention prediction model(IIPM) is proposed, which can quantitatively predict the intensity of interactive intention by the visual features of face orientation, social distance and facial expression in the actual social environment. Based on this model, humanoid robots can make autonomous decisions and select interactive person to carry out interactive tasks reasonably. Finally, the prediction accuracy of the IIPM is proved by single-person and multi-person experiments, which provides an effective and accurate solution for natural human-robot interaction (HRI).
- N. Mavridis. "A review of verbal and non-verbal human–robot interactive communication." Robotics and Autonomous Systems 63 (2015): 22-35. https://doi.org/10.1016/j.robot.2014.09.031.Google ScholarDigital Library
- O. Mubin, "An image based non-verbal behaviour analysis of HRI." International Conference on Social Robotics. Springer, Cham, 2017. https://doi.org/10.1007/978-3-319-70022-9_3.Google Scholar
- T. Tasaki, "Dynamic communication of humanoid robot with multiple people based on interaction distance." Information and Media Technologies 1.1 (2006): 285-295. https://doi.org/10.11185/imt.1.285Google Scholar
- R. Bednarik, H. Vrzakova, and M.l Hradis. 2012. What do you want to do next: a novel approach for intent prediction in gaze-based interaction. In Proceedings of the Symposium on Eye Tracking Research and Applications (ETRA '12). Association for Computing Machinery, New York, NY, USA, 83–90. https://doi.org/10.1145/2168556.2168569.Google ScholarDigital Library
- W. Wang, “Human computer interaction:Intention recognition based on EEG and eyet racking.” Acta Aeronautica et Astronautica Sinica(in Chinese) 42.02(2021):292-302. https://doi:10.7527/S10006893.2020.24290.Google Scholar
- S. Ren, (2019) Research on Interactive Intent Recognition Based on Facial Expression and Line of Sight Direction. In: Li J., Wang S., Qin S., Li X., Wang S. (eds) Advanced Data Mining and Applications. ADMA 2019. Lecture Notes in Computer Science, vol 11888. Springer, Cham. https://doi.org/10.1007/978-3-030-35231-8_31Google Scholar
- T. Baltrusaitis, A. Zadeh, Y. C. Lim and L. Morency, "OpenFace 2.0: Facial Behavior Analysis Toolkit," 2018 13th IEEE International Conference on Automatic Face & Gesture Recognition (FG 2018), 2018, pp. 59-66. https://doi: 10.1109/FG.2018.00019.Google Scholar
- T. L. Saaty, "Decision making with the analytic hierarchy process." International journal of services sciences 1.1 (2008): 83-98. https://doi.org/10.1504/IJSSci.2008.01759Google Scholar
- A. Zadeh, Y. C. Lim, T. Baltrušaitis and L. Morency, "Convolutional Experts Constrained Local Model for 3D Facial Landmark Detection," 2017 IEEE International Conference on Computer Vision Workshops (ICCVW), 2017, pp. 2519-2528. https://doi: 10.1109/ICCVW.2017.296.Google Scholar
- E. T. Hall., "Proxemics [and comments and replies]." Current anthropology 9.2/3 (1968): 83-108. https://doi: 10.1086/200975.Google ScholarCross Ref
- S. M. Fiore, "Toward understanding social cues and signals in human–robot interaction: effects of robot gaze and proxemic behavior." Frontiers in psychology 4 (2013): 859. https://doi: 10.3389/fpsyg.2013.00859Google ScholarCross Ref
- X. T. Truong, T. D. Ngo. "Social interactive intention prediction and categorization." ICRA 2019 Workshop on MoRobAE-Mobile Robot Assistants for the Elderly, Montreal Canada, May 20-24. 2019.Google Scholar
- J. -E. Nurmi, (1996), Optimistic, Approach-oriented, and Avoidance Strategies in Social Situations: Three Studies on Loneliness and Peer Relationships. Eur. J. Pers., 10: 201-219. https://doi.org/10.1002/(SICI)1099-0984(199609)10:3<201::AID-PER257>3.0.CO;2-#.Google Scholar
- Ekman, Paul, and Erika L. Rosenberg, eds. What the face reveals: Basic and applied studies of spontaneous expression using the Facial Action Coding System (FACS). Oxford University Press, USA, 1997.Google Scholar
- Y. -I. Tian, T. Kanade and J. F. Cohn, "Recognizing action units for facial expression analysis," in IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 23, no. 2, pp. 97-115, Feb. 2001. https://doi:10.1109/34.908962.Google ScholarDigital Library
Index Terms
- Interactive Intention Prediction Model for Humanoid Robot Based on Visual Features
Recommendations
Development of a humanoid robot
This study presents design methodologies, specifications and control strategies for vision-guided object grasping for the developed humanoid robot, Cheng-kung Humanoid RobotIc System (CHRIS). The humanoid robot constructed herein comprises mainly a ...
Biomimetic application of desert ant visual navigation for mobile robot docking with weighted landmarks
Previous work has shown that honeybees use a snapshot model to determine a local vector to find their way home. A simpler, average landmark vector model has since been proposed for biologically-inspired mobile robot homing. Previously, the authors have ...
Vision-Based Humanoid Robot Navigation in a Featureless Environment
MCPR 2015: Proceedings of the 7th Mexican Conference on Pattern Recognition - Volume 9116One of the most basic tasks for any autonomous mobile robot is that of safely navigating from one point to another e.g. service robots should be able to find their way in different kinds of environments. Typically, vision is used to find landmarks in ...
Comments