Abstract
The growing deployment of robots in social contexts implies the need to model their behaviour as social agents. In this context, the way a robot approaches a user and eventually engages in an interaction is a crucial aspect to take into account for the acceptance of these tools. In this work, we explore how the approaching policy and gaze behaviours can influence the perceived intention to interact before the interaction starts. The conducted user study highlights the importance of the robot’s gaze behaviour when approaching a human with respect to its approaching behaviour. In particular, if the robot moves in the surroundings of a human, even not straightforward in their direction, but locks the gaze at them, the intention to interact is recognised clearer and faster with respect to the direct approaching of the user but with an adverse gaze.
This work has been supported by the European Union’s Horizon 2020 research and innovation programme under the Marie Skłodowska-Curie grant agreement No 955778.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Admoni, H., Scassellati, B.: Social eye gaze in human-robot interaction: a review. J. Hum.-Rob. Interact. 6(1), 25–63 (2017)
Avelino, J., Garcia-Marques, L., Ventura, R., Bernardino, A.: Break the ice: a survey on socially aware engagement for human-robot first encounters. Int. J. Social Rob. 13(8), 1851–1877 (2021)
Ba, S.O., Odobez, J.M.: Recognizing visual focus of attention from head pose in natural meetings. IEEE Trans. Syst. Man Cybern. Part B (Cybern.) 39(1), 16–33 (2008)
Belpaeme, T., et al.: Multimodal child-robot interaction: building social bonds. J. Hum.-Robot Interact. 1(2) (2012)
Ben-Youssef, A., Clavel, C., Essid, S., Bilac, M., Chamoux, M., Lim, A.: Ue-hri: a new dataset for the study of user engagement in spontaneous human-robot interactions. In: Proceedings of the 19th ACM International Conference on Multimodal Interaction, pp. 464–472 (2017)
Breazeal, C., Dautenhahn, K., Kanda, T.: Social robotics. In: Springer Handbook of Robotics, pp. 1935–1972 (2016)
Faul, F., Erdfelder, E., Lang, A.G., Buchner, A.: G* power 3: a flexible statistical power analysis program for the social, behavioral, and biomedical sciences. Behav. Res. Methods 39(2), 175–191 (2007)
Hall, E.T.: The Hidden Dimension, vol. 609. Anchor (1966)
Heenan, B., Greenberg, S., Aghel-Manesh, S., Sharlin, E.: Designing social greetings in human robot interaction. In: Proceedings of the 2014 Conference on Designing Interactive Systems, pp. 855–864 (2014)
Hong, A., et al.: A multimodal emotional human-robot interaction architecture for social robots engaged in bidirectional communication. IEEE Trans. Cybern. 51(12), 5954–5968 (2021)
Kendon, A.: Conducting Interaction: Patterns of Behavior in Focused Encounters, vol. 7. CUP Archive (1990)
MacArthur, K.R., Stowers, K., Hancock, P.A.: Human-robot interaction: proximity and speed-slowly back away from the robot! In: Advances in Human Factors in Robots and Unmanned Systems, pp. 365–374. Springer, Heidelberg (2017). https://doi.org/10.1007/978-3-319-41959-6_30
Maniscalco, U., Storniolo, P., Messina, A.: Bidirectional multi-modal signs of checking human-robot engagement and interaction. Int. J. Social Rob., 1–15 (2022)
Neggers, M.M., Cuijpers, R.H., Ruijten, P.A., Ijsselsteijn, W.A.: Determining shape and size of personal space of a human when passed by a robot. Int. J. Social Rob. 14(2), 561–572 (2022)
Normoyle, A., Badler, J.B., Fan, T., Badler, N.I., Cassol, V.J., Musse, S.R.: Evaluating perceived trust from procedurally animated gaze. In: Proceedings of Motion on Games, pp. 141–148 (2013)
Oertel, C., et al.: Engagement in human-agent interaction: an overview. Front. Rob. AI 7, 92 (2020)
Patompak, P., Jeong, S., Nilkhamhang, I., Chong, N.Y.: Learning proxemics for personalized human-robot social interaction. Int. J. Social Rob. 12(1), 267–280 (2020)
Pereira, A., Oertel, C., Fermoselle, L., Mendelson, J., Gustafson, J.: Effects of different interaction contexts when evaluating gaze models in hri. In: Proceedings of the 2020 ACM/IEEE International Conference on Human-Robot Interaction, pp. 131–139 (2020)
Rossi, S., Ercolano, G., Raggioli, L., Savino, E., Ruocco, M.: The disappearing robot: an analysis of disengagement and distraction during non-interactive tasks. In: 2018 27th IEEE International Symposium on Robot and Human Interactive Communication (RO-MAN), pp. 522–527 (2018)
Rossi, S., Staffa, M., Bove, L., Capasso, R., Ercolano, G.: User’s personality and activity influence on hri comfortable distances. In: Social Robotics, pp. 167–177. Springer, Cham (2017). https://doi.org/10.1007/978-3-319-70022-9_17
Sanghvi, J., Castellano, G., Leite, I., Pereira, A., McOwan, P.W., Paiva, A.: Automatic analysis of affective postures and body motion to detect engagement with a game companion. In: Proceedings of the 6th International Conference on Human-Robot Interaction, pp. 305–312 (2011)
Satake, S., Kanda, T., Glas, D.F., Imai, M., Ishiguro, H., Hagita, N.: How to approach humans? strategies for social robots to initiate interaction. In: Proceedings of the 4th ACM/IEEE International Conference on Human Robot Interaction, pp. 109–116 (2009)
Sidner, C.L., Lee, C., Kidd, C.D., Lesh, N., Rich, C.: Explorations in engagement for humans and robots. Artif. Intell. 166(1), 140–164 (2005)
Staffa, M., Gregorio, M.D., Giordano, M., Rossi, S.: Can you follow that guy? In: 22th European Symposium on Artificial Neural Networks, ESANN 2014, Bruges, Belgium, 23–25 April 2014 (2014)
Sternberg, R.J., Smith, C.: Social intelligence and decoding skills in nonverbal communication. Social Cogn. 3(2), 168 (1985)
Strazdas, D., Hintz, J., Felßberg, A.M., Al-Hamadi, A.: Robots and wizards: an investigation into natural human-robot interaction. IEEE Access 8, 207635–207642 (2020)
Syrdal, D.S., Dautenhahn, K., Walters, M.L., Koay, K.L.: Sharing spaces with robots in a home scenario-anthropomorphic attributions and their effect on proxemic expectations and evaluations in a live hri trial. In: AAAI Fall Symposium: AI in Eldercare: New Solutions to Old Problems, pp. 116–123 (2008)
Vigni, F., Knoop, E., Prattichizzo, D., Malvezzi, M.: The role of closed-loop hand control in handshaking interactions. IEEE Rob. Autom. Lett. 4(2), 878–885 (2019)
Zhang, Y., Beskow, J., Kjellström, H.: Look but don’t stare: mutual gaze interaction in social robots. In: International Conference on Social Robotics, pp. 556–566. Springer, Heidelberg (2017). https://doi.org/10.1007/978-3-319-70022-9_55
Złotowski, J., Proudfoot, D., Yogeeswaran, K., Bartneck, C.: Anthropomorphism: opportunities and challenges in human-robot interaction. Int. J. Social Rob. 7(3), 347–360 (2015)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2022 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this paper
Cite this paper
Vigni, F., Rossi, S. (2022). Exploring Non-verbal Strategies for Initiating an HRI. In: Cavallo, F., et al. Social Robotics. ICSR 2022. Lecture Notes in Computer Science(), vol 13817. Springer, Cham. https://doi.org/10.1007/978-3-031-24667-8_25
Download citation
DOI: https://doi.org/10.1007/978-3-031-24667-8_25
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-24666-1
Online ISBN: 978-3-031-24667-8
eBook Packages: Computer ScienceComputer Science (R0)