Abstract
In Human-Robot Interaction (HRI), an important quantity of work has been done to investigate the reaction of people toward expressive robots. However, the large variability of available expression modalities (e.g., gaze, gestures, speech modulation) can make comparison between results difficult. We believe that developing a common taxonomy to describe these modalities would contribute to the standardization of HRI experiments. This paper proposes the first version of a classification system based on an analysis of humanoid robots commonly seen and used in HRI studies. Features from the face of robots are discussed in terms of functional and non-functional dimensions, and a short-hand notation is developed to describe these features.
Keywords
These keywords were added by machine and not by the authors. This process is experimental and the keywords may be updated as the learning algorithm improves.
This is a preview of subscription content, log in via an institution.
Buying options
Tax calculation will be finalised at checkout
Purchases are for personal use only
Learn about institutional subscriptionsReferences
Blow, M., Dautenhahn, K., Appleby, A., Nehaniv, C.L., Lee, D.C.: Perception of robot smiles and dimensions for human-robot interaction design. In: Proceedings of the 15th IEEE International Symposium on Robot and Human Interactive Communication, pp. 469–474 (2006)
Satake, S., Kanda, T., Glas, D.F., Imai, M., Ishiguro, H., Hagita, N.: How to approach humans? strategies for social robots to initiate interaction. In: Proceedings of the ACM/IEEE International Conference on Human-Robot Interaction, pp. 109–116 (2009)
Kato, Y., Kanda, T., Ishiguro, H.: May i help you? design of human-like polite approaching behavior. In: Proceedings of the Tenth Annual ACM/IEEE International Conference on Human-Robot Interaction, pp. 35–42 (2015)
Takayama, L., Pantofaru, C.: Influences on proxemic behaviors in human-robot interaction. In: Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems, pp. 5495–5502 (2009)
Sorostinean, M., Ferland, F., Dang, T.-H.-H., Tapus, A.: Motion-oriented attention for a social gaze robot behavior. In: Beetz, M., Johnston, B., Williams, M.-A. (eds.) ICSR 2014. LNCS (LNAI), vol. 8755, pp. 310–319. Springer, Heidelberg (2014). doi:10.1007/978-3-319-11973-1_32
Andrist, S., Mutlu, B., Tapus, A.: Look like me: matching robot personality via gaze to increase motivation. In: Proceedings of the 33rd Annual ACM Conference on Human Factors in Computing Systems, pp. 3603–3612 (2015)
Goetz, J., Kiesler, S., Powers, A.: Matching robot appearance and behavior to tasks to improve human-robot cooperation. In: Proceedings of the 12th IEEE International Workshop on Robot and Human Interactive Communication, pp. 55–60 (2003)
Or, C.K., Duffy, V.G., Cheung, C.C.: Perception of safe robot idle time in virtual reality and real industrial environments. Int. J. Ind. Ergon. 39(5), 807–812 (2009)
Robotics, S.: Who is nao? https://www.ald.softbankrobotics.com/en/cool-robots/nao
Garage, W.: PR2 overview. http://www.willowgarage.com/pages/pr2/overview
von Zitzewitz, J., Boesch, P.M., Wolf, P., Riener, R.: Quantifying the human likeness of a humanoid robot. Int. J. Soc. Robot. 5(2), 263–276 (2013)
Mori, M., MacDorman, K.F., Kageki, N.: The uncanny valley [from the field]. Robot. Autom. Mag. 19(2), 98–100 (2012)
Kedzierski, J., Muszynski, R., Zoll, C., Oleksy, A., Frontkiewicz, M.: Emys–emotive head of a social robot. Int. J. Social Robot. 5(2), 237–249 (2013)
Ribeiro, T., Paiva, A.: The illusion of robotic life: principles and practices of animation for robots. In: Proceedings of the Seventh Annual ACM/IEEE International Conference on Human-Robot Interaction, pp. 383–390 (2012)
Shayganfar, M., Rich, C., Sidner, C.L.: A design methodology for expressing emotion on robot faces. In: Proceedings of the IEEE/RSJ International Conference on Robots and Systems, pp. 4577–4583 (2012)
Beck, A., Hiolle, A., Mazel, A., Cañamero, L.: Interpretation of emotional body language displayed by robots. In: Proceedings of the 3rd International Workshop on Affective Interaction in Natural Environments, pp. 37–42 (2010)
Cohen, I., Looije, R., Neerincx, M.A.: Child’s recognition of emotions in robot’s face and body. In: Proceedings of the 6th International Conference on Human-robot Interaction, pp. 123–124 (2011)
Haring, M., Bee, N., Andre, E.: Creation and evaluation of emotion expression with body movement, sound and eye color for humanoid robots. In: Proceedings of the 20th IEEE International Symposium on Robot and Human Interactive Communication, pp. 204–209 (2011)
Jung, M.F., Lee, J.J., DePalma, N., Adalgeirsson, S.O., Hinds, P.J., Breazeal, C.: Engaging robots: easing complex human-robot teamwork using backchanneling. In: Proceedings of the Conference on Computer Supported Cooperative Work, pp. 1555–1566 (2013)
Strait, M., Canning, C., Scheutz, M.: Let me tell you! investigating the effects of robot communication strategies in advice-giving situations based on robot appearance, interaction modality and distance. In: Proceedings of the ACM/IEEE International Conference on Human-Robot Interaction, pp. 479–486 (2014)
Huang, C.M., Mutlu, B.: Modeling and evaluating narrative gestures for humanlike robots. In: Robotics: Science and Systems (2013)
Muto, Y., Takasugi, S., Yamamoto, T., Miyake, Y.: Timing control of utterance and gesture in interaction between human and humanoid robot. In: The 18th IEEE International Symposium on Robot and Human Interactive Communication, pp. 1022–1028 (2009)
Karg, M., Samadani, A.A., Gorbet, R., Kuhnlenz, K., Hoey, J., Kulic, D.: Body movements for affective expression: a survey of automatic recognition and generation. IEEE Trans. Affect. Comput. 4(4), 341–359 (2013)
Honda: ASIMO specifications. http://asimo.honda.com/asimo-specs/
Robotics, R.: Baxter technical specifications. http://www.rethinkrobotics.com/baxter/tech-specs/
Ferland, F., Létourneau, D., Aumont, A., Frémy, J., Legault, M.A., Lauria, M., Michaud, F.: Natural interaction design of a humanoid robot. J. Hum. Rob. Interact. 1(2), 14–29 (2012)
Rich, C., Sidner, C.L.: Robots and avatars as hosts, advisors, companions, and jesters. AI Mag. 30(1), 29 (2009)
Carter, K., Scheutz, M., Schermerhorn, P.: A humanoid-robotic replica in USARsim for HRI experiments. In: IROS Workshop on Robots, Games, and Research (2009)
Robotics, P.A.L.: REEM: Full-size humanoid service robot. http://pal-robotics.com/en/products/reem/
Metta, G., Natale, L., Nori, F., Sandini, G., Vernon, D., Fadiga, L., Von Hofsten, C., Rosander, K., Lopes, M., Santos-Victor, J., et al.: The iCub humanoid robot: an open-systems platform for research in cognitive development. Neural Netw. 23(8), 1125–1134 (2010)
Moshkina, L., Trickett, S., Trafton, J.G.: Social engagement in public places: a tale of one robot. In: Proceedings of the 2014 ACM/IEEE International Conference on Human-Robot Interaction, pp. 382–389 (2014)
Dietrich, A., Bussmann, K., Petit, F., Kotyczka, P., Ott, C., Lohmann, B., Albu-Schäffer, A.: Whole-body impedance control of wheeled mobile manipulators. Autonomous Robots: Special Issue on Whole-Body Control of Contacts and Dynamics for Humanoid Robots, May 2015
Mitsubishi Heavy Industries: Wakamaru. http://www.mhi-global.com/products/detail/wakamaru_technology.html (available on archive.org)
IEEE Spectrum: Robosoft unveils kompai robot to assist elderly, disabled. http://spectrum.ieee.org/automaton/robotics/medical-robots/robosoft-kompai-robot-assist-elderly-disabled
IEEE Spectrum: Care-O-bot 4 is the robot servant we all want but probably can’t afford. http://spectrum.ieee.org/automaton/robotics/home-robots/care-o-bot-4-mobile-manipulator
Ekman, P., Friesen, W.V.: Facial action coding system: A technique for the measurement of facial movement. Consulting Psychologists Press (1978)
Coulson, M.: Attributing emotion to static body postures: recognition accuracy, confusions, and viewpoint dependence. J. Nonverbal Behav. 28(2), 117–139 (2004)
Acknowledgements
This work has been partially funded by EU Horizon2020 ENRICHME project grant agreement no. 643691C.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2016 Springer International Publishing AG
About this paper
Cite this paper
Ferland, F., Tapus, A. (2016). Functional and Non-functional Expressive Dimensions: Classification of the Expressiveness of Humanoid Robots. In: Agah, A., Cabibihan, JJ., Howard, A., Salichs, M., He, H. (eds) Social Robotics. ICSR 2016. Lecture Notes in Computer Science(), vol 9979. Springer, Cham. https://doi.org/10.1007/978-3-319-47437-3_35
Download citation
DOI: https://doi.org/10.1007/978-3-319-47437-3_35
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-47436-6
Online ISBN: 978-3-319-47437-3
eBook Packages: Computer ScienceComputer Science (R0)