skip to main content
10.1145/3679318.3685351acmotherconferencesArticle/Chapter ViewAbstractPublication PagesnordichiConference Proceedingsconference-collections
research-article

Let Me Explain What I Did or What I Would Have Done: An Empirical Study on the Effects of Explanations and Person-Likeness on Trust in and Understanding of Algorithms

Published: 13 October 2024 Publication History

Abstract

AI-based systems become ubiquitous but stay opaque. To promote informed usage, explanations of how systems compute outcomes emerged as a promising strategy (XAI). Moreover, there is a trend of equipping technologies with human-like features (e.g., natural language, social roles) to enhance the system's person-likeness, thereby influencing users' understanding and trust. The present empirical study (N = 762) investigates the effects and interplay of explanation types and person-likeness on understanding of and trust in AI-systems. It is designed as 2 (contrastive 'why not' explanations vs. comparative 'why' explanation) x 2 (self-referencing vs. neutral formulation) between-subject study with a no explanation-baseline condition. Results demonstrate positive effects of explanations on people's understanding, whereby contrastive explanations – despite being perceived as more complex – led to significantly higher factual understanding compared to the comparative explanations. Person-likeness had no effect on trust or understanding which challenges current trends of enhancing human-like features of technologies.

References

[1]
Or Biran and Courtenay Cotton. 2017. Explanation and justification in machine learning: A survey. In IJCAI ‘17 Workshop on Explainable AI (XAI), 8–13.
[2]
Ruth M. J. Byrne. 2019. Counterfactuals in explainable artificial intelligence (XAI): Evidence from human reasoning. In Proceedings of the 28th International Joint Conference on Artificial Intelligence - IJCAI ‘19, 6276–6282.
[3]
Martin M. Chemers, Li-tze Hu, and Ben F. Garcia. 2001. Academic self-efficacy and first year college student performance and adjustment. Journal of Educational Psychology 93, 1, 55–64.
[4]
Henriette Cramer, Vanessa Evers, Satyan Ramlal, Maarten van Someren, Lloyd Rutledge, Natalia Stash, Lora Aroyo, and Bob Wielinga. 2008. The effects of transparency on trust in and acceptance of a content-based art recommender. User Modeling and User-Adapted Interaction 18, 5, 455–496.
[5]
Richard A. Crosby and William L. Yarber. 2001. Perceived versus actual knowledge about correct condom use among U.S. adolescents: Results from a national study. Journal of Adolescent Health 28, 5, 415–420.
[6]
Maartje De Graaf and Bertram F. Malle. 2017. How people explain action (and autonomous intelligent systems should too). In 2017 AAAI Fall Symposium Series. AI Access Foundation, 19–26.
[7]
Derek Doran, Sarah Schulz, and Tarek R. Besold. 2017. What does explainable AI really mean? A new conceptualization of perspectives.
[8]
Jens Edlund, Joakim Gustafson, Mattias Heldner, and Anna Hjalmarsson. 2008. Towards human-like spoken dialogue systems. Speech Communication 50, 8-9, 630–645.
[9]
Upol Ehsan, Q. V. Liao, Michael Muller, Mark O. Riedl, and Justin D. Weisz. 2021. Expanding explainability: Towards social transparency in AI systems. In Proceedings of the 2021 CHI Conference on Human Factors in Computing Systems - CHI ‘21. ACM, New York, NY, USA, 1–19.
[10]
[10] Friederike Eyssel, Dieta Kuchenbrandt, Simon Bobinger, Laura de Ruiter, and Frank Hegel. 2012. ‘If you sound like me, you must be more human’: On the interplay of robot and user features on human-robot acceptance and anthropomorphism. In Proceedings of the seventh annual ACM/IEEE international conference on Human-Robot Interaction - HRI ‘12. ACM Press, New York, New York, USA, 125–126.
[11]
[11] Kenneth R. Fleischmann and William A. Wallace. 2009. Ensuring transparency in computational modeling. Communications of the ACM 52, 3, 131–134.
[12]
[12] Thomas Franke, Christiane Attig, and Daniel Wessel. 2019. A personal resource for technology interaction: Development and validation of the Affinity for Technology Interaction (ATI) scale. International Journal of Human–Computer Interaction 35, 6, 456–467.
[13]
[13] Gefen and Straub. 2003. Managing user trust in B2C e-services. e-Service Journal 2, 2, 7.
[14]
[14] Gerd Gigerenzer and Wolfgang Gaissmaier. 2011. Heuristic decision making. Annual Review of Psychology 62, 1, 451–482.
[15]
[15] Riccardo Guidotti, Anna Monreale, Fosca Giannotti, Dino Pedreschi, Salvatore Ruggieri, and Franco Turini. 2019. Factual and counterfactual explanations for black box decision making. IEEE Intelligent Systems 34, 6, 14–23.
[16]
[16] David Gunning, Mark Stefik, Jaesik Choi, Tim Miller, Simone Stumpf, and Guang-Zhong Yang. 2019. XAI - Explainable artificial intelligence. Science Robotics 4, 37.
[17]
[17] Hani Hagras. 2018. Toward human-understandable, explainable AI. Computer 51, 9, 28–36.
[18]
[18] Joseph F. Hair, William C. Black, Barry J. Babin, and Rolph E. Anderson. 2009. Multivariate Data Analysis. (7th ed.). Pearson Prentice Hall.
[19]
[19] Tracy Harwood and Tony Garry. 2017. Internet of Things: Understanding trust in techno-service systems. Journal of Service Management 28, 3, 442–475.
[20]
[20] Robert R. Hoffman, Tim Miller, Shane T. Mueller, Gary Klein, and William J. Clancey. 2018. Explaining explanation, part 4: A deep dive on deep nets. IEEE Intelligent Systems 33, 3, 87–95.
[21]
[21] Robert R. Hoffman, Shane T. Mueller, Gary Klein, and Jordan Litman. 2018. Metrics for explainable AI: Challenges and prospects.
[22]
[22] Regina Jucks, Gesa A. Linnemann, Franziska M. Thon, and Maria Zimmermann. 2016. Trust the words: Insights into the role of language in trust building in a digitalized world. In Trust and Communication in a Digitized World, Bernd Blöbaum, Ed. Progress in IS. Springer, Cham, 225–237.
[23]
[23] K. Karrer, C. Glaser, C. Clemens, and C. Bruder. 2009. Technikaffinität erfassen: der Fragebogen TA-EG [Measuring technical affinity - the questionnaire TA-EG]. Der Mensch im Mittelpunkt Technischer Systeme, 8, 196–201.
[24]
[24] S. Kiesler. 2005. Fostering common ground in human-robot interaction. In Proceedings of the 2005 IEEE International Workshop on Robot and Human Interactive Communication - ROMAN ‘05. IEEE, 729–734.
[25]
[25] Nicole C. Krämer, Astrid von der Pütten, and Sabrina Eimler. 2012. Human-agent and human-robot interaction theory: Similarities to and differences from human-human interaction. In Human-Computer Interaction: The Agency Perspective, Marielba Zacarias and José V. de Oliveira, Eds. Studies in Computational Intelligence. Springer, Berlin, Heidelberg, 215–240.
[26]
[26] Justin Kruger and David Dunning. 1999. Unskilled and unaware of it: How difficulties in recognizing one's own incompetence lead to inflated self-assessments. Journal of Personality and Social Psychology 77, 6, 1121–1134.
[27]
[27] Philipp Kulms and Stefan Kopp. 2019. More human-likeness, more trust? In Proceedings of Mensch und Computer 2019. ACM, New York, NY, USA, 31–42.
[28]
[28] Pete Ladwig, Kajsa E. Dalrymple, Dominique Brossard, Dietram A. Scheufele, and Elizabeth A. Corley. 2012. Perceived familiarity or factual knowledge? Comparing operationalizations of scientific understanding. Science and Public Policy 39, 6, 761–774.
[29]
[29] Nancy Lankton, D. H. McKnight, and John Tripp. 2015. Technology, humanness, and trust: Rethinking trust in technology. Journal of the Association for Information Systems 16, 10, 880–918.
[30]
[30] John D. Lee and Katrina A. See. 2004. Trust in automation: Designing for appropriate reliance. Human Factors 46, 1, 50–80.
[31]
[31] Cedric A. Lehmann, Christine B. Haubitz, Andreas Fügener, and Ulrich W. Thonemann. 2020. Keep it mystic? The effects of algorithm transparency on the use of advice. In Forty-First International Conference on Information Systems - ICIS ‘20, 1–9.
[32]
[32] James W. Lichtenberg and William A. Moffitt. 1994. The effect of predicate matching on perceived understanding and factual recall. Journal of Counseling & Development 72, 5, 544–548.
[33]
[33] Brian Y. Lim, Anind K. Dey, and Daniel Avrahami. 2009. Why and why not explanations improve the intelligibility of context-aware intelligent systems. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems - CHI ‘09. ACM, New York, NY, USA, 2119–2128.
[34]
[34] Xiao Ma, Jeffery T. Hancock, Kenneth Lim Mingjie, and Mor Naaman. 2017. Self-disclosure and perceived trustworthiness of Airbnb host profiles. In Proceedings of the 2017 ACM Conference on Computer Supported Cooperative Work and Social Computing - CSCW ‘17. ACM, New York, NY, USA, 2397–2409.
[35]
[35] Maria Madsen and Shirley Gregor. 2000. Measuring human-computer trust. In Proceedings of the 11th Australasian Conference on Information Systems. Australasian Association for Information Systems, 6–8.
[36]
[36] Roger C. Mayer, James H. Davis, and F. D. Schoorman. 1995. An integrative model of organizational trust. The Academy of Management Review 20, 3, 709.
[37]
[37] D. H. McKnight, Michelle Carter, Jason B. Thatcher, and Paul F. Clay. 2011. Trust in a specific technology. ACM Transactions on Management Information Systems 2, 2, 1–25.
[38]
[38] D. H. McKnight, Vivek Choudhury, and Charles Kacmar. 2002. Developing and validating trust measures for e-commerce: An integrative typology. Information Systems Research 13, 3, 334–359.
[39]
[39] Tim Miller. 2019. Explanation in artificial intelligence: Insights from the social sciences. Artificial Intelligence 267, 1–38.
[40]
[40] Brent Mittelstadt, Chris Russell, and Sandra Wachter. 2019. Explaining explanations in AI. In Proceedings of the Conference on Fairness, Accountability, and Transparency - FAT ‘19. ACM, New York, NY, USA, 279–288.
[41]
[41] Janet Morahan-Martin and Phyllis Schumacher. 2007. Attitudinal and experiential predictors of technological expertise. Computers in Human Behavior 23, 5, 2230–2239.
[42]
[42] Clifford Nass and Scott Brave. 2005. Wired for Speech: How Voice Activates and Advances the Human–Computer Relationship. The MIT Press, Cambridge, MA.
[43]
[43] Marc Prensky. 2001. Digital natives, digital immigrants part 2: Do they really think differently? On the Horizon 9, 6, 1–6.
[44]
[44] Jianhong Qu, Ronggang Zhou, Liming Zou, Yanyan Sun, and Min Zhao. 2020. The effect of personal pronouns on users’ emotional experience in voice interaction. In Human-Computer Interaction. Multimodal and Natural Interaction. HCII 2020. Lecture Notes in Computer Science, Masaaki Kurosu, Ed., 12182. Springer, Cham, 234–243.
[45]
[45] Byron Reeves and Clifford Nass. 1996. The Media Equation: How People Treat Computers, Television, and New Media Like Real People and Places. Cambridge University Press, Cambridge, UK.
[46]
[46] Marco T. Ribeiro, Sameer Singh, and Carlos Guestrin. 2016. “Why should I trust you?”: Explaining the predictions of any classifier. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining - KDD ‘16. ACM, New York, NY, USA, 1135–1144.
[47]
[47] Wojciech Samek, Grégoire Montavon, Andrea Vedaldi, Lars K. Hansen, and Klaus-Robert Müller. 2019. Explainable AI: Interpreting, Explaining and Visualizing Deep Learning, 11700. Springer, Cham.
[48]
[48] Laura Schelenz, Avi Segal, and Kobi Gal. 2020. Applying Transparency in Artificial Intelligence based Personalization Systems.
[49]
[49] Donghee Shin. 2021. The effects of explainability and causability on perception, trust, and acceptance: Implications for explainable AI. International Journal of Human-Computer Studies 146, 102551.
[50]
[50] Dan Sperber, Fabrice Clément, Christophe Heintz, Olivier Mascaro, Hugo Mercier, Gloria Origgi, and Deirdre Wilson. 2010. Epistemic vigilance. Mind & Language 25, 4, 359–393.
[51]
[51] Leona Y.-F. Su, Michael A. Cacciatore, Dietram A. Scheufele, Dominique Brossard, and Michael A. Xenos. 2014. Inequalities in scientific understanding: Differentiating between factual and perceived knowledge gaps. Science Communication 36, 3, 352–378.
[52]
[52] Kashvi Taunk, Sanjukta De, Srishti Verma, and Aleena Swetapadma. 2019. A brief review of nearest neighbor algorithm for learning and classification. In International Conference on Intelligent Computing and Control Systems - ICCS ‘19. IEEE, 1255–1260.
[53]
[53] Michelle M. van Pinxteren, Mark Pluymaekers, and Jos G. Lemmink. 2020. Human-like communication in conversational agents: A literature review and research agenda. Journal of Service Management 31, 2, 203–225.
[54]
[54] Sahil Verma, John Dickerson, and Keegan Hines. 2020. Counterfactual explanations for machine learning: A review.
[55]
[55] Ewart J. de Visser, Samuel S. Monfort, Ryan McKendrick, Melissa A. B. Smith, Patrick E. McKnight, Frank Krueger, and Raja Parasuraman. 2016. Almost human: Anthropomorphism increases trust resilience in cognitive agents. Journal of Experimental Psychology: Applied 22, 3, 331–349.
[56]
[56] Jacob Weisberg, Dov Te'eni, and Limor Arman. 2011. Past purchase and intention to purchase in e‐commerce. Internet Research 21, 1, 82–96.
[57]
[57] Torsten Wilholt. 2013. Epistemic trust in science. The British Journal for the Philosophy of Science 64, 2, 233–253.
[58]
[58] Robert H. Wortham and Andreas Theodorou. 2017. Robot transparency, trust and utility. Connection Science 29, 3, 242–248.
[59]
[59] Maria Zimmermann and Regina Jucks. 2018. Investigating the role of communication for information seekers’ trust-related evaluations of health videos on the web: Content analysis, survey data, and experiment. Interactive Journal of Medical Research 7, 2, e10282.

Cited By

View all
  • (2024)Do Perceived Benefits Compensate for Low Provider Trustworthiness in Disclosure Decisions?Journal of Media Psychology10.1027/1864-1105/a000440Online publication date: 16-Jul-2024

Index Terms

  1. Let Me Explain What I Did or What I Would Have Done: An Empirical Study on the Effects of Explanations and Person-Likeness on Trust in and Understanding of Algorithms

    Recommendations

    Comments

    Information & Contributors

    Information

    Published In

    cover image ACM Other conferences
    NordiCHI '24: Proceedings of the 13th Nordic Conference on Human-Computer Interaction
    October 2024
    1236 pages
    ISBN:9798400709661
    DOI:10.1145/3679318
    Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than the author(s) must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected].

    Publisher

    Association for Computing Machinery

    New York, NY, United States

    Publication History

    Published: 13 October 2024

    Permissions

    Request permissions for this article.

    Check for updates

    Qualifiers

    • Research-article
    • Research
    • Refereed limited

    Funding Sources

    Conference

    NordiCHI 2024

    Acceptance Rates

    Overall Acceptance Rate 379 of 1,572 submissions, 24%

    Contributors

    Other Metrics

    Bibliometrics & Citations

    Bibliometrics

    Article Metrics

    • Downloads (Last 12 months)85
    • Downloads (Last 6 weeks)32
    Reflects downloads up to 03 Mar 2025

    Other Metrics

    Citations

    Cited By

    View all
    • (2024)Do Perceived Benefits Compensate for Low Provider Trustworthiness in Disclosure Decisions?Journal of Media Psychology10.1027/1864-1105/a000440Online publication date: 16-Jul-2024

    View Options

    Login options

    View options

    PDF

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader

    HTML Format

    View this article in HTML Format.

    HTML Format

    Figures

    Tables

    Media

    Share

    Share

    Share this Publication link

    Share on social media