Abstract
To enable socially situated human-robot interaction, a robot must both understand and control proxemics—the social use of space—to employ communication mechanisms analogous to those used by humans. In this work, we considered how proxemic behavior is influenced by human speech and gesture production, and how this impacts robot speech and gesture recognition in face-to-face social interactions. We conducted a data collection to model these factors conditioned on distance. This resulting models of pose, speech, and gesture were consistent with related work in human-human interactions, but were inconsistent with related work in human-human interactions—participants in our data collection pos itioned themselves much farther away than has been observed in related work. These models have been integrated into a situated autonomous proxemic robot controller, in which the robot selects interagent pose parameters to maximize its expectation to recognize natural human speech and body gestures during an interaction. This work contributes to the understanding of the underlying per-cultural processes that govern human proxemic behavior, and has implications for the development of robust proxemic controllers for sociable and socially assistive robots situated in complex interactions (e.g., with multiple people or individuals with hearing/visual impairments) and environments (e.g., in which there is loud noise, reverberation, low lighting, or visual occlusion).
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Notes
- 1.
- 2.
Our current work considers only dyadic interactions (i.e., interactions between two agents); however, our framework is extensible [9], providing a principled approach (supported by related work [2]) for spatially situated interactions between any number of sociable agents by maximizing how each of them will produce and perceive speech and gestures, as well as any other social signals not considered by this work.
- 3.
- 4.
Recall that participants were “strangers” or, at most, “acquaintances”.
- 5.
References
Tapus, A., Matarić, M., Scassellati, B.: The grand challenges in socially assistive robotics. IEEE Robot. Autom. Mag. 14(1), 35–42 (2007)
Hall, E.T.: The Hidden Dimension. Doubleday Company, Chicago (1966)
Hall, E.T.: American Anthropology Association. Handbook for Proxemic Research. Washington (1974)
Torta, E., Cuijpers, R.H., Juola, J.F., van der Pol, D.: Design of robust robotic proxemic behaviour. In: Proceedings of the Third International Conference on Social Robotics, ICSR’11, pp. 21–30 (2011)
Kuzuoka, H., Suzuki, Y., Yamashita, J., Yamazaki, K.: Reconfiguring spatial formation arrangement by robot body orientation. In: HRI. Osaka (2010)
Walters, M., Dautenhahn, K., Boekhorst, R., Koay, K., Syrdal, D., Nehaniv, C.: An empirical framework for human-robot proxemics. In: New Frontiers in Human-Robot Interaction, pp. 144–149. Edinburgh (2009)
Mumm, J., Mutlu, B.: Human-robot proxemics: Physical and psychological distancing in human-robot interaction. In: HRI, pp. 331–338. Lausanne (2011)
Takayama, L., Pantofaru, C.: Influences on proxemic behaviors in human-robot interaction. In: IEEE/RSJ International Conference on Intelligent Robots and Systems, IROS’09, pp. 5495–5502 (2009)
Mead, R., Matarić, M.J.: A probabilistic framework for autonomous proxemic control in situated and mobile human-robot interaction. In: 7th ACM/IEEE International Conference on Human-Robot Interaction, pp. 193–194. Boston (2012)
Hall, E.T.: The Silent Language. Doubleday Company, New York (1959)
Hall, E.: A system for notation of proxemic behavior. Am. Anthropol. 65, 1003–1026 (1963)
Mehrabian, A.: Nonverbal Communication. Aldine Transcation, Piscataway (1972)
Vasquez, D., Stein, P., Rios-Martinez, J., Escobedo, A., Spalanzani, A., Laugier, C.: Human aware navigation for assistive robotics. In: Proceedings of the Thirteenth International Symposium on Experimental Robotics, ISER’12. Québec City (2012)
Trautman, P., Krause, A.: Unfreezing the robot: navigation in dense, interacting crowds. In: IROS. Taipei (2010)
Satake, S., Kanda, T., Glas, D.F., Imai, M., Ishiguro, H., Hagita, N.: How to approach humans?: strategies for social robots to initiate interaction. In: HRI, pp. 109–116 (2009)
Feil-Seifer, D., Matarić, M.: Automated detection and classification of positive versus negative robot interactions with children with autism using distance-based features. In: HRI’11, pp. 323–330. Lausanne (2011)
Oosterhout, T., Visser, A.: A visual method for robot proxemics measurements. In: HRI Workshop on Metrics for Human-Robot Interaction. Amsterdam (2008)
Jones, S., Aiello, J.: Proxemic behavior of black and white first-, third-, and fifth-grade children. J. Personal. Soc. Psychol. 25(1), 21–27 (1973)
Mead, R., Atrash, A., Matarić, M.J.: Proxemic feature recognition for interactive robots: automating metrics from the social sciences. In: International Conference on Social Robotics, pp. 52–61. Amsterdam (2011)
Mead, R., Atrash, A., Matarić, M.J.: Automated proxemic feature extraction and behavior recognition: applications in human-robot interaction. Int. J. Soc. Robot. 5(3), 367–378 (2013)
Mead, R., Atrash, A., Matarić, M.J.: Recognition of spatial dynamics for predicting social interaction. In: 6th ACM/IEEE International Conference on Human-Robot Interaction, pp. 201–202. Lausanne (2011)
Feil-Seifer, D., Matarić, M.: People-aware navigation for goal-oriented behavior involving a human partner. In: IEEE International Conference on Development and Learning, ICDL’11, vol. 2, pp. 1–6. Frankfurt Am Main, Germany (2011)
Mead, R., Atrash, A., Matarić, M.J.: Representations of proxemic behavior for human-machine interaction. In: NordiCHI 2012 Workshop on Proxemics in Human-Computer Interaction, Copenhagen (2012)
McNeill, D.: Hand and Mind: What Gestures Reveal about Thought. Chicago University Press, Chicago (1992)
Brooks, A.G., Arkin, R.C.: Behavioral overlays for non-verbal communication expression on a humanoid robot. Auton. Robot. 22(1), 55–74 (2007)
Kriz, S., Anderson, G., Trafton, J.G.: Robot-directed speech: using language to assess first-time users’ conceptualizations of a robot. In: Proceedings of the 5th ACM/IEEE International Conference on Human-Robot Interaction, HRI’10, pp. 267–274. IEEE Press (2010)
Kendon, A.: Conducting Interaction—Patterns of Behavior in Focused Encounters. Cambridge University Press, New York (1990)
Ozyurek, A.: Do speakers design their co-speech gestures for their addresees? the effects of addressee location on representational gestures. J. Mem. Lang. 46(4), 688–704 (2002)
Wöelfel, M., McDonough, J.: Distant Speech Recognition. Wiley, West Sussex (2009)
Gardner, B., Martin, K.: HRTF measurements of a kemar dummy-head microphone. Technical Report 280. MIT Media Lab Perceptual Computing, Boston (1994)
Chu, W., Warnock, A.: Detailed Directivity of Sound Fields Around Human Talkers. Research Report. Institute for Research in Construction, Canada (2002)
Acknowledgments
This work is supported in part by an NSF Graduate Research Fellowship, the NSF National Robotics Initiative (IIS-1208500), NSF IIS-1117279 and CNS-0709296 grants, and the PR2 Beta Program.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2016 Springer International Publishing Switzerland
About this chapter
Cite this chapter
Mead, R., Matarić, M.J. (2016). Perceptual Models of Human-Robot Proxemics. In: Hsieh, M., Khatib, O., Kumar, V. (eds) Experimental Robotics. Springer Tracts in Advanced Robotics, vol 109. Springer, Cham. https://doi.org/10.1007/978-3-319-23778-7_18
Download citation
DOI: https://doi.org/10.1007/978-3-319-23778-7_18
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-23777-0
Online ISBN: 978-3-319-23778-7
eBook Packages: EngineeringEngineering (R0)