skip to main content
10.1145/3491102.3517565acmconferencesArticle/Chapter ViewAbstractPublication PageschiConference Proceedingsconference-collections
research-article
Public Access

Owning Mistakes Sincerely: Strategies for Mitigating AI Errors

Published:29 April 2022Publication History

ABSTRACT

Interactive AI systems such as voice assistants are bound to make errors because of imperfect sensing and reasoning. Prior human-AI interaction research has illustrated the importance of various strategies for error mitigation in repairing the perception of an AI following a breakdown in service. These strategies include explanations, monetary rewards, and apologies. This paper extends prior work on error mitigation by exploring how different methods of apology conveyance may affect people’s perceptions of AI agents; we report an online study (N=37) that examines how varying the sincerity of an apology and the assignment of blame (on either the agent itself or others) affects participants’ perceptions and experience with erroneous AI agents. We found that agents that openly accepted the blame and apologized sincerely for mistakes were thought to be more intelligent, likeable, and effective in recovering from errors than agents that shifted the blame to others.

Skip Supplemental Material Section

Supplemental Material

3491102.3517565-talk-video.mp4

mp4

21.7 MB

References

  1. Gavin Abercrombie, Amanda Cercas Curry, Mugdha Pandya, and Verena Rieser. 2021. Alexa, Google, Siri: What are Your Pronouns? Gender and Anthropomorphism in the Design and Perception of Conversational Assistants. arXiv preprint arXiv:2106.02578(2021).Google ScholarGoogle Scholar
  2. Karin Aijmer. 2019. ‘Ooh whoops I’m sorry! Teenagers’ use of English apology expressions. Journal of Pragmatics 142 (2019), 258–269.Google ScholarGoogle ScholarCross RefCross Ref
  3. Christoph Bartneck, Dana Kulić, Elizabeth Croft, and Susana Zoghbi. 2009. Measurement instruments for the anthropomorphism, animacy, likeability, perceived intelligence, and perceived safety of robots. International journal of social robotics 1, 1 (2009), 71–81.Google ScholarGoogle ScholarCross RefCross Ref
  4. Dennis Benner, Edona Elshan, Sofia Schöbel, and Andreas Janson. 2021. What do you mean? A Review on Recovery Strategies to Overcome Conversational Breakdowns of Conversational Agents. In International Conference on Information Systems (ICIS).Google ScholarGoogle Scholar
  5. Alison Wood Brooks, Hengchen Dai, and Maurice E Schweitzer. 2014. I’m sorry about the rain! Superfluous apologies demonstrate empathic concern and increase trust. Social Psychological and Personality Science 5, 4 (2014), 467–474.Google ScholarGoogle ScholarCross RefCross Ref
  6. Daniel J Brooks, Momotaz Begum, and Holly A Yanco. 2016. Analysis of reactions towards failures and recovery strategies for autonomous robots. In 2016 25th IEEE International Symposium on Robot and Human Interactive Communication (RO-MAN). IEEE, 487–492.Google ScholarGoogle ScholarDigital LibraryDigital Library
  7. David Cameron, Stevienna de Saille, Emily C Collins, Jonathan M Aitken, Hugo Cheung, Adriel Chua, Ee Jing Loh, and James Law. 2021. The effect of social-cognitive recovery strategies on likability, capability and trust in social robots. Computers in Human Behavior 114 (2021), 106561.Google ScholarGoogle ScholarDigital LibraryDigital Library
  8. James L Heskett Christopher W Hart and Jr W Earl Sasser. 1990. The profitable art of service recovery. (1990), 148–156.Google ScholarGoogle Scholar
  9. Jacob Cohen. 1988. Statistical power analysis for the behavioral sciences. England: Routledge (1988).Google ScholarGoogle Scholar
  10. Filipa Correia, Carla Guerra, Samuel Mascarenhas, Francisco S Melo, and Ana Paiva. 2018. Exploring the impact of fault justification in human-robot trust. In Proceedings of the 17th international conference on autonomous agents and multiagent systems. 507–513.Google ScholarGoogle ScholarDigital LibraryDigital Library
  11. Ryan Fehr and Michele J Gelfand. 2010. When apologies work: How matching apology components to victims’ self-construals facilitates forgiveness. Organizational behavior and human decision processes 113, 1(2010), 37–50.Google ScholarGoogle Scholar
  12. Joel E Fischer, Stuart Reeves, Martin Porcheron, and Rein Ove Sikveland. 2019. Progressivity for voice interface design. In Proceedings of the 1st International Conference on Conversational User Interfaces. 1–8.Google ScholarGoogle ScholarDigital LibraryDigital Library
  13. John Fought. 1972. Erving Goffman, Relations in public: microstudies of the public order. New York: Basic Books, 1971. Pp. xvii 396.Language in Society 1, 2 (1972), 266–271. https://doi.org/10.1017/S0047404500000543Google ScholarGoogle ScholarCross RefCross Ref
  14. Caleb Furlough, Thomas Stokes, and Douglas J Gillan. 2021. Attributing blame to robots: I. The influence of robot autonomy. Human factors 63, 4 (2021), 592–602.Google ScholarGoogle Scholar
  15. Xiang Ge, Dan Li, Daisong Guan, Shihui Xu, Yanyan Sun, and Moli Zhou. 2019. Do smart speakers respond to their errors properly? A study on human-computer dialogue strategy. In International Conference on Human-Computer Interaction. Springer, 440–455.Google ScholarGoogle ScholarDigital LibraryDigital Library
  16. Petra Gieselmann. 2006. Comparing error-handling strategies in human-human and human-robot dialogues. In Proc. 8th Conf. Nat. Language Process.(KONVENS). Konstanz, Germany. 24–31.Google ScholarGoogle Scholar
  17. Trudy Govier and Wilhelm Verwoerd. 2002. The promise and pitfalls of apology. Journal of social philosophy 33, 1 (2002), 67–82.Google ScholarGoogle ScholarCross RefCross Ref
  18. David Griol and José Manuel Molina. 2016. A framework for improving error detection and correction in spoken dialog systems. Soft Computing 20, 11 (2016), 4229–4241.Google ScholarGoogle ScholarDigital LibraryDigital Library
  19. Victoria Groom, Jimmy Chen, Theresa Johnson, F Arda Kara, and Clifford Nass. 2010. Critic, compatriot, or chump?: Responses to robot blame attribution. In 2010 5th ACM/IEEE International Conference on Human-Robot Interaction (HRI). IEEE, 211–217.Google ScholarGoogle ScholarCross RefCross Ref
  20. Yaou Hu, Hyounae Min, and Na Su. 2021. How Sincere is an Apology? Recovery Satisfaction in A Robot Service Failure Context. Journal of Hospitality & Tourism Research(2021), 10963480211011533.Google ScholarGoogle Scholar
  21. Gilhwan Hwang, Jeewon Lee, Cindy Yoonjung Oh, and Joonhwan Lee. 2019. It sounds like a woman: Exploring gender stereotypes in South Korean voice assistants. In Extended Abstracts of the 2019 CHI Conference on Human Factors in Computing Systems. 1–6.Google ScholarGoogle ScholarDigital LibraryDigital Library
  22. Amazon Inc.[n.d.]. Alexa Design Guide. https://developer.amazon.com/en-US/docs/alexa/alexa-design/get-started.html. Accessed: 2021-12-01.Google ScholarGoogle Scholar
  23. Google Inc.[n.d.]. Conversation Design. https://developers.google.com/assistant/conversation-design/welcome. Accessed: 2021-12-01.Google ScholarGoogle Scholar
  24. Jiepu Jiang, Wei Jeng, and Daqing He. 2013. How do users respond to voice input errors? Lexical and phonetic query reformulation in voice search. In Proceedings of the 36th international ACM SIGIR conference on Research and development in information retrieval. 143–152.Google ScholarGoogle ScholarDigital LibraryDigital Library
  25. Ing-Marie Jonsson, Clifford Nass, Jack Endo, Ben Reaves, Helen Harris, Janice Le Ta, Nicholas Chan, and Sean Knapp. 2004. Don’t blame me I am only the Driver: Impact of Blame Attribution on Attitudes and Attention to Driving Task. In CHI’04 extended abstracts on Human factors in computing systems. 1219–1222.Google ScholarGoogle Scholar
  26. Poornima Kaniarasu and Aaron M Steinfeld. 2014. Effects of blame on trust in human robot interaction. In The 23rd IEEE international symposium on robot and human interactive communication. IEEE, 850–855.Google ScholarGoogle ScholarCross RefCross Ref
  27. Peter H Kim, Kurt T Dirks, Cecily D Cooper, and Donald L Ferrin. 2006. When more blame is better than less: The implications of internal vs. external attributions for the repair of trust after a competence-vs. integrity-based trust violation. Organizational behavior and human decision processes 99, 1 (2006), 49–65.Google ScholarGoogle Scholar
  28. Taenyun Kim and Hayeon Song. 2021. How should intelligent agents apologize to restore trust? Interaction effects between anthropomorphism and apology attribution on trust repair. Telematics and Informatics 61 (2021), 101595.Google ScholarGoogle ScholarCross RefCross Ref
  29. Dimosthenis Kontogiorgos, Sanne van Waveren, Olle Wallberg, Andre Pereira, Iolanda Leite, and Joakim Gustafson. 2020. Embodiment effects in interactions with failing robots. In Proceedings of the 2020 CHI conference on human factors in computing systems. 1–14.Google ScholarGoogle ScholarDigital LibraryDigital Library
  30. Anastasia Kuzminykh, Jenny Sun, Nivetha Govindaraju, Jeff Avery, and Edward Lank. 2020. Genie in the bottle: Anthropomorphized perceptions of conversational agents. In Proceedings of the 2020 CHI Conference on Human Factors in Computing Systems. 1–13.Google ScholarGoogle ScholarDigital LibraryDigital Library
  31. Min Kyung Lee, Sara Kiesler, Jodi Forlizzi, Siddhartha Srinivasa, and Paul Rybski. 2010. Gracefully mitigating breakdowns in robotic services. In 2010 5th ACM/IEEE International Conference on Human-Robot Interaction (HRI). IEEE, 203–210.Google ScholarGoogle Scholar
  32. Xuying Leo and Young Eun Huh. 2020. Who gets the blame for service failures? Attribution of responsibility toward robot versus human service providers and service firms. Computers in Human Behavior 113 (2020), 106520.Google ScholarGoogle ScholarCross RefCross Ref
  33. Gina-Anne Levow. 1998. Characterizing and recognizing spoken corrections in human-computer dialogue. In 36th Annual Meeting of the Association for Computational Linguistics and 17th International Conference on Computational Linguistics, Volume 1. 736–742.Google ScholarGoogle Scholar
  34. Roderick JA Little. 1988. A test of missing completely at random for multivariate data with missing values. Journal of the American statistical Association 83, 404(1988), 1198–1202.Google ScholarGoogle ScholarCross RefCross Ref
  35. James G Maxham III and Richard G Netemeyer. 2003. Firms reap what they sow: the effects of shared values and perceived organizational justice on customers’ evaluations of complaint handling. Journal of Marketing 67, 1 (2003), 46–62.Google ScholarGoogle ScholarCross RefCross Ref
  36. Steven J Migacz, Suiwen Zou, and James F Petrick. 2018. The “terminal” effects of service failure on airlines: Examining service recovery with justice theory. Journal of Travel Research 57, 1 (2018), 83–98.Google ScholarGoogle ScholarCross RefCross Ref
  37. Nicole Mirnig, Gerald Stollnberger, Markus Miksch, Susanne Stadler, Manuel Giuliani, and Manfred Tscheligi. 2017. To err is robot: How humans assess and act toward an erroneous social robot. Frontiers in Robotics and AI 4 (2017), 21.Google ScholarGoogle ScholarCross RefCross Ref
  38. Wade J Mitchell, Chin-Chang Ho, Himalaya Patel, and Karl F MacDorman. 2011. Does social desirability bias favor humans? Explicit–implicit evaluations of synthesized speech support a new HCI model of impression management. Computers in Human Behavior 27, 1 (2011), 402–412.Google ScholarGoogle ScholarDigital LibraryDigital Library
  39. Flora Moon. [n.d.]. Sound of Text. https://soundoftext.com/. Accessed: 2021-08-01.Google ScholarGoogle Scholar
  40. Bob Moore and Raphael Arar. [n.d.]. Conversation design Guidelines. https://conversational-ux.mybluemix.net/design/conversational-ux/. Accessed: 2021-12-01.Google ScholarGoogle Scholar
  41. Chelsea Myers, Anushay Furqan, Jessica Nebolsky, Karina Caro, and Jichen Zhu. 2018. Patterns for how users overcome obstacles in voice user interfaces. In Proceedings of the 2018 CHI Conference on Human Factors in Computing Systems. 1–7.Google ScholarGoogle ScholarDigital LibraryDigital Library
  42. Cathy Pearl. 2016. Designing voice user interfaces: Principles of conversational experiences. ” O’Reilly Media, Inc.”.Google ScholarGoogle Scholar
  43. Valentina Pitardi and Hannah R Marriott. 2021. Alexa, she’s not human but… Unveiling the drivers of consumers’ trust in voice-based artificial intelligence. Psychology & Marketing 38, 4 (2021), 626–642.Google ScholarGoogle ScholarCross RefCross Ref
  44. Martin Porcheron, Joel E Fischer, Stuart Reeves, and Sarah Sharples. 2018. Voice interfaces in everyday life. In proceedings of the 2018 CHI conference on human factors in computing systems. 1–12.Google ScholarGoogle ScholarDigital LibraryDigital Library
  45. Martin Porcheron, Joel E Fischer, and Sarah Sharples. 2017. ” Do Animals Have Accents?” Talking with Agents in Multi-Party Conversation. In Proceedings of the 2017 ACM conference on computer supported cooperative work and social computing. 207–219.Google ScholarGoogle Scholar
  46. Marianne Promberger and Jonathan Baron. 2006. Do patients trust computers?Journal of Behavioral Decision Making 19, 5 (2006), 455–468.Google ScholarGoogle Scholar
  47. Holger Roschk and Susanne Kaiser. 2013. The nature of an apology: An experimental study on how to apologize after a service failure. Marketing Letters 24, 3 (2013), 293–309.Google ScholarGoogle ScholarCross RefCross Ref
  48. Donald B Rubin. 2004. Multiple imputation for nonresponse in surveys. Vol. 81. John Wiley & Sons.Google ScholarGoogle Scholar
  49. Jagdish N Sheth and CH Kellstadt. 1994. A normative model of retaining customer satisfaction. Gamma News JournalJuly-August (1994), 4–7.Google ScholarGoogle Scholar
  50. Richard Spreng, Gilbert Harrell, and Robert Mackoy. 1995. Service Recovery: Impact on Satisfaction and Intentions. Journal of Services Marketing 9 (03 1995), 15–23. https://doi.org/10.1108/08876049510079853Google ScholarGoogle ScholarCross RefCross Ref
  51. Stephen S Tax, Stephen W Brown, and Murali Chandrashekaran. 1998. Customer evaluations of service complaint experiences: implications for relationship marketing. Journal of marketing 62, 2 (1998), 60–76.Google ScholarGoogle ScholarCross RefCross Ref
  52. Joost R van Ginkel, Marielle Linting, Ralph CA Rippe, and Anja van der Voort. 2020. Rebutting existing misconceptions about multiple imputation as a method for handling missing data. Journal of personality assessment 102, 3 (2020), 297–308.Google ScholarGoogle ScholarCross RefCross Ref

Index Terms

  1. Owning Mistakes Sincerely: Strategies for Mitigating AI Errors

          Recommendations

          Comments

          Login options

          Check if you have access through your login credentials or your institution to get full access on this article.

          Sign in
          • Published in

            cover image ACM Conferences
            CHI '22: Proceedings of the 2022 CHI Conference on Human Factors in Computing Systems
            April 2022
            10459 pages
            ISBN:9781450391573
            DOI:10.1145/3491102

            Copyright © 2022 ACM

            Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than the author(s) must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected].

            Publisher

            Association for Computing Machinery

            New York, NY, United States

            Publication History

            • Published: 29 April 2022

            Permissions

            Request permissions about this article.

            Request Permissions

            Check for updates

            Qualifiers

            • research-article
            • Research
            • Refereed limited

            Acceptance Rates

            Overall Acceptance Rate6,199of26,314submissions,24%

          PDF Format

          View or Download as a PDF file.

          PDF

          eReader

          View online with eReader.

          eReader

          HTML Format

          View this article in HTML Format .

          View HTML Format