Abstract
In this paper, we investigate pedestrian interaction with a large autonomous robot that clears snow from the sidewalk. Through a virtual reality (VR) based user study, simulating different robot behaviors, we report on perceptions of encountering a potentially dangerous robot on the sidewalk. Overall, participants considered their actions in VR to be representative of their real-world actions. However, we note that VR headsets are not able to reproduce the high dynamic range required to realistically reproduce the high-intensity warning lights and sounds associated with close proximity to a large industrial machine. Participants expressed concern about interrupting the robot’s work, and that the robot’s safety-driven behavior should not delay passing by it on the sidewalk.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Abdrabou, Y., et al.: Understanding shoulder surfer behavior and attack patterns using virtual reality. In: Proceedings of the 2022 International Conference on Advanced Visual Interfaces, pp. 1–9 (2022)
Bartneck, C., Forlizzi, J.: A design-centred framework for social human-robot interaction. In: RO-MAN 2004, 13th IEEE International Workshop on Robot and Human Interactive Communication (IEEE Catalog No. 04TH8759), pp. 591–594. IEEE (2004)
Bogue, R.: Domestic robots: has their time finally come? Ind. Robot Int. J. (2017)
Chadalavada, R.T., Andreasson, H., Krug, R., Lilienthal, A.J.: That’s on my mind! Robot to human intention communication through on-board projection on shared floor space. In: 2015 European Conference on Mobile Robots (ECMR), pp. 1–6. IEEE (2015)
Coovert, M.D., Lee, T., Shindev, I., Sun, Y.: Spatial augmented reality as a method for a mobile robot to communicate intended movement. Comput. Hum. Behav. 34, 241–248 (2014)
Dancu, A., et al.: Gesture bike: examining projection surfaces and turn signal systems for urban cycling. In: Proceedings of the 2015 International Conference on Interactive Tabletops and Surfaces, pp. 151–159 (2015)
Dey, D., et al.: Taming the EHMI jungle: a classification taxonomy to guide, compare, and assess the design principles of automated vehicles’ external human-machine interfaces. Transport. Res. Interdisc. Perspect. 7, 100174 (2020)
Dey, D., Habibovic, A., Pfleging, B., Martens, M., Terken, J.: Color and animation preferences for a light band EHMI in interactions between automated vehicles and pedestrians. In: Proceedings of the 2020 CHI Conference on Human Factors in Computing Systems. CHI ’20, pp. 1–13. Association for Computing Machinery, New York, NY, USA (2020). https://doi.org/10.1145/3313831.3376325
Galar, D., Kumar, U., Seneviratne, D.: Robots, Drones, UAVs and UGVs for Operation and Maintenance. CRC Press, Boca Raton (2020)
Gielniak, M.J., Thomaz, A.L.: Generating anticipation in robot motion. In: 2011 RO-MAN, pp. 449–454. IEEE (2011)
Han, Z., Parrillo, J., Wilkinson, A., Yanco, H.A., Williams, T.: Projecting robot navigation paths: hardware and software for projected ar. arXiv preprint arXiv:2112.05172 (2021)
Hart, J., et al.: Using human-inspired signals to disambiguate navigational intentions. In: Wagner, A.R., et al. (eds.) ICSR 2020. LNCS (LNAI), vol. 12483, pp. 320–331. Springer, Cham (2020). https://doi.org/10.1007/978-3-030-62056-1_27
Holländer, K., Colley, A., Mai, C., Häkkilä, J., Alt, F., Pfleging, B.: Investigating the influence of external car displays on pedestrians’ crossing behavior in virtual reality. In: Proceedings of the 21st International Conference on Human-Computer Interaction with Mobile Devices and Services. MobileHCI ’19. Association for Computing Machinery, New York, NY, USA (2019). https://doi.org/10.1145/3338286.3340138
Kassem, K., Ungerböck, T., Wintersberger, P., Michahelles, F.: What is happening behind the wall? Towards a better understanding of a hidden robot’s intent by multimodal cues. Proc. ACM Hum.-Comput. Interact. 6(MHCI), 1–19 (2022)
Kleygrewe, L., Hutter, R.V., Koedijk, M., Oudejans, R.R.: Virtual reality training for police officers: a comparison of training responses in VR and real-life training. Police Pract. Res. 1–20 (2023)
Lee, A., Toombs, A.L.: Robots on campus: understanding public perception of robots using social media. In: Conference Companion Publication of the 2020 on Computer Supported Cooperative Work and Social Computing, pp. 305–309 (2020)
Lee, Y.M., et al.: Learning to interpret novel EHMI: the effect of vehicle kinematics and EHMI familiarity on pedestrian’ crossing behavior. J. Safety Res. 80, 270–280 (2022)
Liu, C., Ishi, C.T., Ishiguro, H., Hagita, N.: Generation of nodding, head tilting and eye gazing for human-robot dialogue interaction. In: 2012 7th ACM/IEEE International Conference on Human-Robot Interaction (HRI), pp. 285–292. IEEE (2012)
Mäkelä, V., et al.: Virtual field studies: conducting studies on public displays in virtual reality. In: Proceedings of the 2020 CHI Conference on Human Factors in Computing Systems, pp. 1–15 (2020)
Mathis, F., et al.: Remote XR studies: the golden future of HCI research. In: Proceedings of the CHI 2021 Workshop on XR Remote Research (2021). https://mat.qmul.ac.uk/events/xr-chi-2021
Moore, D., Currano, R., Strack, G.E., Sirkin, D.: The case for implicit external human-machine interfaces for autonomous vehicles. In: Proceedings of the 11th International Conference on Automotive User Interfaces and Interactive Vehicular Applications. AutomotiveUI ’19, pp. 295–307. Association for Computing Machinery, New York, NY, USA (2019). https://doi.org/10.1145/3342197.3345320
Pizzi, G., Scarpi, D., Pichierri, M., Vannucci, V.: Virtual reality, real reactions?: Comparing consumers’ perceptions and shopping orientation across physical and virtual-reality retail stores. Comput. Hum. Behav. 96, 1–12 (2019)
Pörtner, A., Schröder, L., Rasch, R., Sprute, D., Hoffmann, M., König, M.: The power of color: a study on the effective use of colored light in human-robot interaction. In: 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), pp. 3395–3402. IEEE (2018)
Rivu, R., et al.: Remote VR studies-a framework for running virtual reality studies remotely via participant-owned HMDs. arXiv preprint arXiv:2102.11207 (2021)
Sagnier, C., Loup-Escande, E., Valléry, G.: Effects of gender and prior experience in immersive user experience with virtual reality. In: Ahram, T., Falcão, C. (eds.) AHFE 2019. AISC, vol. 972, pp. 305–314. Springer, Cham (2020). https://doi.org/10.1007/978-3-030-19135-1_30
Schneider, S., Bengler, K.: Virtually the same? Analysing pedestrian behaviour by means of virtual reality. Transport. Res. F: Traffic Psychol. Behav. 68, 231–256 (2020)
Vázquez, M., Carter, E.J., McDorman, B., Forlizzi, J., Steinfeld, A., Hudson, S.E.: Towards robot autonomy in group conversations: Understanding the effects of body orientation and gaze. In: 2017 12th ACM/IEEE International Conference on Human-Robot Interaction (HRI), pp. 42–52. IEEE (2017)
Vice: Robot dog attack (2023). https://www.vice.com/en/article/3ad9zj/dollar15000-robot-dog-walks-30-seconds-on-brisbane-street-before-someone-punts-it-in-the-head. Accessed 26 Jan 2023
Voit, A., Mayer, S., Schwind, V., Henze, N.: Online, VR, AR, lab, and in-situ: comparison of research methods to evaluate smart artifacts. In: Proceedings of the 2019 CHI Conference on Human Factors in Computing Systems, pp. 1–12 (2019)
Zhang, J., Vinkhuyzen, E., Cefkin, M.: Evaluation of an autonomous vehicle external communication system concept: a survey study. In: Stanton, N.A. (ed.) AHFE 2017. AISC, vol. 597, pp. 650–661. Springer, Cham (2018). https://doi.org/10.1007/978-3-319-60441-1_63
Acknowledgments
This research has been supported by the Lapland Robotics and Innovation in Lapland through Design and Art (ILO) projects, funded by the European Regional Development Fund (ERDF).
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
1 Electronic supplementary material
Below is the link to the electronic supplementary material.
Supplementary material 1 (mp4 23857 KB)
Rights and permissions
Copyright information
© 2023 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this paper
Cite this paper
Colley, A., Tiitto, M., Pfleging, B., Häkkilä, J. (2023). Pedestrian Interaction with a Snow Clearing Robot. In: Abdelnour Nocera, J., Kristín Lárusdóttir, M., Petrie, H., Piccinno, A., Winckler, M. (eds) Human-Computer Interaction – INTERACT 2023. INTERACT 2023. Lecture Notes in Computer Science, vol 14143. Springer, Cham. https://doi.org/10.1007/978-3-031-42283-6_1
Download citation
DOI: https://doi.org/10.1007/978-3-031-42283-6_1
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-42282-9
Online ISBN: 978-3-031-42283-6
eBook Packages: Computer ScienceComputer Science (R0)