Abstract
Humans are very apt at reading emotional signals in other humans and even artificial agents, which raises the question of whether artificial agents need to be emotionally intelligent to ensure effective social interactions. For artificial agents without emotional intelligence might generate behavior that is misinterpreted, unexpected, and confusing to humans, violating human expectations and possibly causing emotional harm. Surprisingly, there is a dearth of investigations aimed at understanding the extent to which artificial agents need emotional intelligence for successful interactions. Here, we present the first study in the perception of emotional intelligence (EI) in robots vs. humans. The objective was to determine whether people viewed robots as more or less emotionally intelligent when exhibiting similar behaviors as humans, and to investigate which verbal and nonverbal communication methods were most crucial for human observational judgments. Study participants were shown a scene in which either a robot or a human behaved with either high or low empathy, and then they were asked to evaluate the agent’s emotional intelligence and trustworthiness. The results showed that participants could consistently distinguish the high EI condition from the low EI condition regardless of the variations in which communication methods were observed, and that whether the agent was a robot or human had no effect on the perception. We also found that relative to low EI high EI conditions led to greater trust in the agent, which implies that we must design robots to be emotionally intelligent if we wish for users to trust them.
Preview
Unable to display preview. Download preview PDF.
References
Ashkanasy, N., Daus, C.: Emotion in the workplace: The new challenge for managers. The Academy of Management Executive 16(1), 76–86 (2002)
Baylor, A., Kim, Y.: Simulating instructional roles through pedagogical agents. International Journal of Artificial Intelligence in Education 15(2), 95–115 (2005)
Bickmore, T., Picard, R.: Establishing and maintaining long-term human-computer relationships. ACM Transactions on Computer-Human Interaction (TOCHI) 12(2), 293–327 (2005)
Brackett, M., Rivers, S., Salovey, P.: Emotional intelligence: Implications for personal, social, academic, and workplace success. Social and Personality Psychology Compass 5, 88–103 (2011)
Caruso, D.: Emotional intelligence scale (in preparation)
Clarke, N.: Emotional intelligence and its relationship to transformational leadership and key project manager competences. Project Management Journal 41(2), 5–20 (2010)
Fariselli, L., Freedman, J., Ghini, M.M., Valentini, F.: Stress, emotional intelligence, and performance in healthcare (retrieved 07/19/2017)
Fasola, J., Mataric, M.: Using socially assistive human–robot interaction to motivate physical exercise for older adults. Proceedings of the IEEE 100, 2512–2526 (2012)
Hagelskamp, C., Brackett, M., Rivers, S., Salovey, P.: Improving classroom quality with the ruler approach to social and emotional learning: Proximal and distal outcomes. American Journal of Community Psychology 51, 530–543 (2013)
Hancock, P., Billings, D., Schaefer, K., Chen, J., Visser, E.D., Parasuraman, R.: A meta-analysis of factors affecting trust in human-robot interaction. Human Factors: The Journal of the Human Factors and Ergonomics Society 53(5), 517–527 (2011)
Kanda, T., Shiomi, M., Miyashita, Z., Ishiguro, H., Hagita, N.: An affective guide robot in a shopping mall. In: 4th ACM/IEEE International Conference on Human-Robot Interaction (HRI), pp. 173–180, March 2009
Lee, J.D., See, K.A.: Trust in automation: Designing for appropriate reliance. Human Factors: The Journal of the Human Factors and Ergonomics Society 46(1), 50–80 (2004)
Lohani, M., Stokes, C., Dashan, N., McCoy, M., Rivers, S.E., Bailey, C.A.: A framework for human-agent social systems: the role of non-technical factors in operation success. In: Proceedings of Advances in Human Factors in Robots and Unmanned Systems, pp. 137–148 (2017)
Lohani, M., Stokes, C., McCoy, M., Bailey, C., Joshi, A., Rivers, S.: Perceived role of physiological sensors impacts trust and reliance on robots. In: Proceedings of 25th IEEE International Symposium on Robot and Human Interactive Communication, pp. 513–518 (2016)
Lohani, M., Stokes, C., McCoy, M., Bailey, C., Rivers, S.: Social interaction moderates human-robot trust-reliance relationship and improves stress coping. In: Proceedings of 11th ACM/IEEE International Conference on Human-Robot Interaction, pp. 471–472 (2016)
Lohani, M., Stokes, C., Oden, K., Frazier, S., Landers, K., Craven, P., Lawton, D., McCoy, M., Macannuco, D.: A framework for human-machine social systems: The influence of non-technical factors on trust and stress appraisal. ACM Transactions in Interactive Intelligence Systems (forthcoming)
Lopes, P., Grewal, D., Kadis, J., Gall, M., Salovey, P.: Evidence that emotional intelligence is related to job performance and affect and attitudes at work. Psicothema 18, 132–138 (2006)
Mayer, R.C., Davis, J.H.: The effect of the performance appraisal system on trust for management: A field quasi-experiment. Journal of Applied Psychology 84(1), 123 (1999)
Mayer, R.C., Davis, J.H., Schoorman, F.D.: An integrative model of organizational trust. Academy of Management Review 20, 709–734 (1995)
Momeni, N.: The relation between managers’ emotional intelligence and the organizational climate they create. Public Personnel Management 38(2), 35–48 (2009)
Picard, R.: Toward machines with emotional intelligence. In: ICINCO (Invited Speakers), pp. 29–30 (2004)
Pontier, M., Siddiqui, G.F.: A virtual therapist that responds empathically to your answers. In: Prendinger, H., Lester, J., Ishizuka, M. (eds.) IVA 2008. LNCS, vol. 5208, pp. 417–425. Springer, Heidelberg (2008). doi:10.1007/978-3-540-85483-8_42
Rivers, S.E., Brackett, M.A., Reyes, M.R., Elbertson, N.A., Salovey, P.: Improving the social and emotional climate of classrooms: A clustered randomized controlled trial testing the ruler approach. Prevention Science 14(1), 77–87 (2013)
Salovey, P., Mayer, J.: Emotional intelligence. Imagination, Cognition and Personality 9(3), 185–211 (1990)
Schaefer, K., Billings, D., Szalma, J., Adams, J., Sanders, T., Chen, J., Hancock, P.: A meta-analysis of factors influencing the development of trust in automation: Implications for human-robot interaction. DTIC Document, Tech. Rep (2014)
de Visser, E., Parasuraman, R.: Adaptive aiding of human-robot teaming effects of imperfect automation on performance, trust, and workload. Journal of Cognitive Engineering and Decision Making 5(2), 209–231 (2011)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2017 Springer International Publishing AG
About this paper
Cite this paper
Fan, L., Scheutz, M., Lohani, M., McCoy, M., Stokes, C. (2017). Do We Need Emotionally Intelligent Artificial Agents? First Results of Human Perceptions of Emotional Intelligence in Humans Compared to Robots. In: Beskow, J., Peters, C., Castellano, G., O'Sullivan, C., Leite, I., Kopp, S. (eds) Intelligent Virtual Agents. IVA 2017. Lecture Notes in Computer Science(), vol 10498. Springer, Cham. https://doi.org/10.1007/978-3-319-67401-8_15
Download citation
DOI: https://doi.org/10.1007/978-3-319-67401-8_15
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-67400-1
Online ISBN: 978-3-319-67401-8
eBook Packages: Computer ScienceComputer Science (R0)