skip to main content
10.1145/3313831.3376624acmconferencesArticle/Chapter ViewAbstractPublication PageschiConference Proceedingsconference-collections
research-article
Public Access

No Explainability without Accountability: An Empirical Study of Explanations and Feedback in Interactive ML

Published: 23 April 2020 Publication History

Abstract

Automatically generated explanations of how machine learning (ML) models reason can help users understand and accept them. However, explanations can have unintended consequences: promoting over-reliance or undermining trust. This paper investigates how explanations shape users' perceptions of ML models with or without the ability to provide feedback to them: (1) does revealing model flaws increase users' desire to "fix" them; (2) does providing explanations cause users to believe - wrongly - that models are introspective, and will thus improve over time. Through two controlled experiments - varying model quality - we show how the combination of explanations and user feedback impacted perceptions, such as frustration and expectations of model improvement. Explanations without opportunity for feedback were frustrating with a lower quality model, while interactions between explanation and feedback for the higher quality model suggest that detailed feedback should not be requested without explanation. Users expected model correction, regardless of whether they provided feedback or received explanations.

References

[1]
Julius Adebayo, Justin Gilmer, Michael Muelly, Ian Goodfellow, Moritz Hardt, and Been Kim. 2018. Sanity Checks for Saliency Maps. In Proceedings of Advances in Neural Information Processing Systems.
[2]
Jae Wook Ahn, Peter Brusilovsky, Jonathan Grady, Daqing He, and Sue Yeon Syn. 2007. Open User Profiles for Adaptive News Systems: Help or Harm?. In Proceedings of the World Wide Web Conference.
[3]
David Alvarez-Melis and Tommi S. Jaakkola. 2018. Towards Robust Interpretability with Self-explaining Neural Networks. In Proceedings of Advances in Neural Information Processing Systems.
[4]
Saleema Amershi, Maya Cakmak, William Bradley Knox, and Todd Kulesza. 2014. Power to the People: The Role of Humans in Interactive Machine Learning. AI Magazine (2014), 105--120.
[5]
Saleema Amershi, James Fogarty, Ashish Kapoor, and Desney Tan. 2010. Examining Multiple Potential Models in End-user Interactive Concept Learning. In International Conference on Human Factors in Computing Systems.
[6]
Saleema Amershi, James Fogarty, and Daniel Weld. 2012. ReGroup: Interactive Machine learning for On-demand Group Creation in Social Networks. In International Conference on Human Factors in Computing Systems.
[7]
Maria-Florina Balcan and Avrim Blum. 2008. Clustering with Interactive Feedback. In International Conference on Algorithmic Learning Theory.
[8]
Gagan Bansal, Besmira Nushi, Ece Kamar, Daniel S Weld, Walter S Lasecki, and Eric Horvitz. 2019. Updates in Human-AI Teams: Understanding and Addressing the Performance/Compatibility Tradeoff. Association for the Advancement of Artificial Intelligence (2019).
[9]
Mustafa Bilgic and Raymond J Mooney. 2005. Explaining Recommendations: Satisfaction vs. Promotion. In Proceedings of Beyond Personalization 2005: A Workshop on the Next Stage of Recommender Systems Research at IUI.
[10]
Or Biran and Kathleen McKeown. 2017. Human-centric Justification of Machine Learning Predictions. In International Joint Conference on Artificial Intelligence.
[11]
Andrea Bunt, Joanna McGrenere, and Cristina Conati. 2007. Understanding the Utility of Rationale in a Mixed-Initiative System for GUI Customization. In International Conference on User Modeling.
[12]
Carrie J. Cai, Jonas Jongejan, and Jess Holbrook. 2019. The Effects of Example-based Explanations in a Machine Learning Interface. In International Conference on Intelligent User Interfaces.
[13]
Oana Maria Camburu, Tim Rocktäschel, Thomas Lukasiewicz, and Phil Blunsom. 2018. E-SNLI: Natural Language Inference with Natural Language Explanations. In Proceedings of Advances in Neural Information Processing Systems.
[14]
Rich Caruana, Yin Lou, Johannes Gehrke, Paul Koch, Marc Sturm, and Noémie Elhadad. 2015. Intelligible Models for Healthcare: Predicting Pneumonia Risk and Hospital 30-day Readmission. In Knowledge Discovery and Data Mining.
[15]
Arjun Chandrasekaran, Viraj Prabhu, Deshraj Yadav, Prithvijit Chattopadhyay, and Devi Parikh. 2018. Do Explanations make VQA Models more Predictable to a Human?. In Proceedings of Empirical Methods in Natural Language Processing.
[16]
Henriette Cramer, Vanessa Evers, Satyan Ramlal, Maarten Van Someren, Lloyd Rutledge, Natalia Stash, Lora Aroyo, and Bob Wielinga. 2008. The Effects of Transparency on Trust in and Acceptance of a Content-based Art Recommender. User Modeling and User-Adapted Interaction (2008).
[17]
Aron Culotta, Trausti Kristjansson, Andrew McCallum, and Paul Viola. 2006. Corrective Feedback and Persistent Learning for Information Extraction. Artificial Intelligence (2006).
[18]
Jonathan Dodge, Q. Vera Liao, Yunfeng Zhang, Rachel K.E. Bellamy, and Casey Dugan. 2019. Explaining Models: An Empirical Study of How Explanations Impact Fairness Judgment. In International Conference on Intelligent User Interfaces.
[19]
Upol Ehsan, Pradyumna Tambwekar, Larry Chan, Brent Harrison, and Mark O. Riedl. 2019. Automated Rationale Generation: A Technique for Explainable AI and its Effects on Human Perceptions. In International Conference on Intelligent User Interfaces.
[20]
Jerry Alan Fails and Dan R Olsen. 2003. Interactive Machine Learning. In International Conference on Intelligent User Interfaces.
[21]
Shi Feng and Jordan Boyd-Graber. 2019. What can AI do for me? Evaluating Machine Learning Interpretations in Cooperative Play. In International Conference on Intelligent User Interfaces.
[22]
Rebecca Fiebrink, Dan Trueman, and Perry R Cook. 2009. A Metainstrument for Interactive, On-the-fly Machine Learning. In Proceedings of New Interfaces for Musical Expression.
[23]
Dimitra Gkatzia, Oliver Lemon, and Verena Rieser. 2016. Natural Language Generation Enhances Human Decision-making with Uncertain Information. In Proceedings of the Association for Computational Linguistics.
[24]
Dorota Glowacka, Tuukka Ruotsalo, Ksenia Konyushkova, Kumaripaba Athukorala, Samuel Kaski, and Giulio Jacucci. 2013. Directing Exploratory Search: Reinforcement Learning from User Interactions with Keywords. In International Conference on Intelligent User Interfaces.
[25]
Bryce Goodman and Seth Flaxman. 2017. European Union Regulations on Algorithmic Decision Making and a -Right to Explanation". AI Magazine (2017).
[26]
Dave Gunning. 2016. Explainable Artificial Intelligence (XAI). (2016). https://www.darpa.mil/program/ explainable-artificial-intelligence
[27]
Jonathan L. Herlocker, Joseph A. Konstan, and John Riedl. 2000. Explaining Collaborative Filtering Recommendations. In Conference on Computer Supported Cooperative Work and Social Computing.
[28]
Kristina Höök. 2000. Steps to Take before Intelligent User Interfaces become Real. Interacting With Computers 12, 4 (2000), 409--426.
[29]
Yuening Hu, Jordan Boyd-Graber, Brianna Satinoff, and Alison Smith. 2014. Interactive Topic Modeling. Machine Learning 95, 3 (6 2014), 423--469.
[30]
Anna Huang. 2008. Similarity Measures for Text Document Clustering. In New Zealand Computer Science Research Student Conference.
[31]
W. Bradley Knox and Peter Stone. 2012. Reinforcement Learning from Human Reward: Discounting in Episodic Tasks. In IEEE International Workshop on Robot and Human Interactive Communication.
[32]
Rafal Kocielnik, Saleema Amershi, and Paul N Bennett. 2019. Will you Accept an Imperfect AI? Exploring Designs for Adjusting End-user Expectations of AI systems. In International Conference on Human Factors in Computing Systems.
[33]
Todd Kulesza, Margaret Burnett, Weng-Keen Wong, and Simone Stumpf. 2015. Principles of Explanatory Debugging to Personalize Interactive Machine Learning. In International Conference on Intelligent User Interfaces.
[34]
Todd Kulesza, Simone Stumpf, Margaret Burnett, and Irwin Kwan. 2012. Tell me More? The Effects of Mental Model Soundness on Personalizing an Intelligent Agent. In International Conference on Human Factors in Computing Systems.
[35]
Todd Kulesza, Simone Stumpf, Margaret Burnett, Sherry Yang, Irwin Kwan, and Weng Keen Wong. 2013. Too Much, too Little, or Just Right? Ways Explanations Impact End Users' Mental Models. In IEEE Symposium on Visual Languages and Human-Centric Computing.
[36]
Isaac Lage, Andrew Slavin Ross, Been Kim, Samuel J. Gershman, and Finale Doshi-Velez. 2018. Human-in-the-loop Interpretability Prior. In Proceedings of Advances in Neural Information Processing Systems.
[37]
Himabindu Lakkaraju, Rich Caruana, Ece Kamar, and Jure Leskovec. 2019. Faithful and Customizable Explanations of Black Box Models. In Conference on AI, Ethics, and Society.
[38]
Ken Lang. 1995. NewsWeeder: Learning to Filter Netnews. In Machine Learning Proceedings.
[39]
Hanseung Lee, Jaeyeon Kihm, Jaegul Choo, John Stasko, and Haesun Park. 2012. iVisClustering: An Interactive Visual Document Clustering via Topic Modeling. Computer Graphics Forum 31 (2012), 1155--1164.
[40]
Tao Lei, Regina Barzilay, and Tommi Jaakkola. 2016. Rationalizing Neural Predictions. In Proceedings of Empirical Methods in Natural Language Processing.
[41]
David D Lewis. 1998. Naive (Bayes) at Forty: The Independence Assumption in Information Retrieval. In Proceedings of European Conference of Machine Learning.
[42]
Brian Lim, Anind Dey, and Daniel Avrahami. 2009. Why and Why Not Explanations Improve the Intelligibility of Context-aware Intelligent Systems. (2009).
[43]
Brian Y. Lim and Anind K. Dey. 2011. Investigating Intelligibility for Uncertain Context-aware Applications. In Proceedings of the international conference on Ubiquitous computing.
[44]
Menaka Narayanan, Emily Chen, Jeffrey He, Been Kim, Sam Gershman, and Finale Doshi-Velez. 2018. How do Humans Understand Explanations from Machine Learning Systems? An Evaluation of the Human-interpretability of Explanation. (2018). http://arxiv.org/abs/1802.00682
[45]
Dong Huk Park, Lisa Anne Hendricks, Zeynep Akata, Anna Rohrbach, Bernt Schiele, Trevor Darrell, and Marcus Rohrbach. 2017. Attentive Explanations: Justifying Decisions and Pointing to the Evidence. In Computer Vision and Pattern Recognition.
[46]
Fabian Pedregosa, Gael Varoquaux, Alexandre Gramfort, Vincent Michel, Bertrand Thirion, Olivier Grisel, Mathieu Blondel, Peter Prettenhofer, Ron Weiss, Vincent Dubourg, Jake Vanderplas, Alexandre Passos, David Cournapeau, Matthieu Brucher, Matthieu Perrot, and Édouard Duchesnay. 2011. scikit-learn: Machine Learning in Python. Journal of Machine Learning Research 12 (2011), 2825--2830.
[47]
Forough Poursabzi-Sangdeh, Daniel G Goldstein, Jake M Hofman, Jennifer Wortman Vaughan, and Hanna Wallach. 2018. Manipulating and Measuring Model Interpretability. (2018). http://arxiv.org/abs/1802.07810
[48]
Pearl Pu and Li Chen. 2006. Trust Building with Explanation Interfaces. In International Conference on Intelligent User Interfaces.
[49]
Hema Raghavan, Omid Madani, and Rosie Jones. 2006. Active Learning with Feedback on both Features and Instances. Journal of Machine Learning Research (2006).
[50]
Paul Resnick, Neophytos Iacovou, Mitesh Suchak, Peter Bergstrom, and John T. Riedl. 1994. GroupLens: An Open Architecture for Collaborative Filtering of Netnews. In Conference on Computer Supported Cooperative Work and Social Computing.
[51]
Marco Tulio Ribeiro, Sameer Singh, and Carlos Guestrin. 2016. Why Should I Trust You? Explaining the Predictions of any Classifier. In Knowledge Discovery and Data Mining.
[52]
Stephanie Rosenthal and Anind K Dey. 2010. Towards Maximizing the Accuracy of Human-labeled Sensor Data. In International Conference on Intelligent User Interfaces.
[53]
William Saunders, Andreas Stuhlmüller, Girish Sastry, and Owain Evans. 2018. Trial without Error: Towards Safe Reinforcement Learning via Human Intervention. In International Joint Conference on Autonomous Agents and Multiagent Systems.
[54]
Philipp Schmidt and Felix Biessmann. 2019. Quantifying Interpretability and Trust in Machine Learning Systems. (2019). http://arxiv.org/abs/1901.08558
[55]
Ramprasaath R. Selvaraju, Michael Cogswell, Abhishek Das, Ramakrishna Vedantam, Devi Parikh, and Dhruv Batra. 2017. Grad-CAM: Visual Explanations from Deep Networks via Gradient-based Localization. In International Conference on Computer Vision.
[56]
Burr Settles. 2010. Active Learning Literature Survey. Machine Learning 15, 2 (2010), 201--221.
[57]
Burr Settles. 2011. Closing the Loop: Fast, Interactive Semi-supervised Annotation with Queries on Features and Instances. In Proceedings of Empirical Methods in Natural Language Processing.
[58]
Zhangzhang Si and Song Chun Zhu. 2013. Learning and-or Templates for Object Recognition and Detection. IEEE Transactions on Pattern Analysis and Machine Intelligence (2013).
[59]
Robert S. Siegler. 2002. Microgenetic Studies of Self-explanation. Cambridge University Press, 31--58.
[60]
Alison Smith, Varun Kumar, Jordan Boyd-Graber, Kevin Seppi, and Leah Findlater. 2018. Closing the Loop: User-centered Design and Evaluation of a Human-in-the-loop Topic Modeling System. In International Conference on Intelligent User Interfaces.
[61]
Kimberly Stowers, Nicholas Kasdaglis, Michael Rupp, Jessie Chen, Daniel Barber, and Michael Barnes. 2017. Insights into Human-agent Teaming: Intelligent Agent Transparency and Uncertainty. In Advances in Intelligent Systems and Computing.
[62]
Simone Stumpf. 2016. Explanations Considered Harmful? User Interactions with Machine Learning Systems. In ACM SIGCHI Workshop on Human-Centered Machine Learning.
[63]
Emily Wall, Soroush Ghorashi, and Gonzalo Ramos. 2019. Using Expert Patterns in Assisted Interactive Machine Learning: A Study in Machine Teaching. In IFIP Conference on Human-Computer Interaction.
[64]
Jacob O. Wobbrock, Leah Findlater, Darren Gergle, and James J. Higgins. 2011. The Aligned Rank Transform for Nonparametric Factorial Analyses Using only ANOVA Procedures. In International Conference on Human Factors in Computing Systems.
[65]
Tongshuang Wu, Daniel S. Weld, and Jeffrey Heer. 2019. Local Decision Pitfalls in Interactive Machine Learning: An Investigation into Feature Selection in Sentiment Analysis. ACM Transactions on Computer-Human Interaction (2019).
[66]
Ming Yin, Jennifer Wortman Vaughan, and Hanna Wallach. 2019. Understanding the Effect of Accuracy on Trust in Machine Learning Models. In International Conference on Human Factors in Computing Systems.

Cited By

View all
  • (2025)Questioning the Robot: Using Human Non-verbal Cues to Estimate the Need for ExplanationsProceedings of the 2025 ACM/IEEE International Conference on Human-Robot Interaction10.5555/3721488.3721577(717-728)Online publication date: 4-Mar-2025
  • (2025)Human-AI collaboration is not very collaborative yet: a taxonomy of interaction patterns in AI-assisted decision making from a systematic reviewFrontiers in Computer Science10.3389/fcomp.2024.15210666Online publication date: 6-Jan-2025
  • (2025)"My Very Subjective Human Interpretation": Domain Expert Perspectives on Navigating the Text Analysis Loop for Topic ModelsProceedings of the ACM on Human-Computer Interaction10.1145/37012019:1(1-30)Online publication date: 10-Jan-2025
  • Show More Cited By

Index Terms

  1. No Explainability without Accountability: An Empirical Study of Explanations and Feedback in Interactive ML

      Recommendations

      Comments

      Information & Contributors

      Information

      Published In

      cover image ACM Conferences
      CHI '20: Proceedings of the 2020 CHI Conference on Human Factors in Computing Systems
      April 2020
      10688 pages
      ISBN:9781450367080
      DOI:10.1145/3313831
      Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

      Sponsors

      Publisher

      Association for Computing Machinery

      New York, NY, United States

      Publication History

      Published: 23 April 2020

      Permissions

      Request permissions for this article.

      Check for updates

      Author Tags

      1. explainable machine learning
      2. interactive machine learning

      Qualifiers

      • Research-article

      Funding Sources

      • NSF
      • ONR

      Conference

      CHI '20
      Sponsor:

      Acceptance Rates

      Overall Acceptance Rate 6,199 of 26,314 submissions, 24%

      Upcoming Conference

      CHI 2025
      ACM CHI Conference on Human Factors in Computing Systems
      April 26 - May 1, 2025
      Yokohama , Japan

      Contributors

      Other Metrics

      Bibliometrics & Citations

      Bibliometrics

      Article Metrics

      • Downloads (Last 12 months)594
      • Downloads (Last 6 weeks)77
      Reflects downloads up to 05 Mar 2025

      Other Metrics

      Citations

      Cited By

      View all
      • (2025)Questioning the Robot: Using Human Non-verbal Cues to Estimate the Need for ExplanationsProceedings of the 2025 ACM/IEEE International Conference on Human-Robot Interaction10.5555/3721488.3721577(717-728)Online publication date: 4-Mar-2025
      • (2025)Human-AI collaboration is not very collaborative yet: a taxonomy of interaction patterns in AI-assisted decision making from a systematic reviewFrontiers in Computer Science10.3389/fcomp.2024.15210666Online publication date: 6-Jan-2025
      • (2025)"My Very Subjective Human Interpretation": Domain Expert Perspectives on Navigating the Text Analysis Loop for Topic ModelsProceedings of the ACM on Human-Computer Interaction10.1145/37012019:1(1-30)Online publication date: 10-Jan-2025
      • (2025)What makes a ‘good’ decision with artificial intelligence? A grounded theory study in paediatric careBMJ Evidence-Based Medicine10.1136/bmjebm-2024-112919(bmjebm-2024-112919)Online publication date: 12-Feb-2025
      • (2025)What large language models know and what people think they knowNature Machine Intelligence10.1038/s42256-024-00976-77:2(221-231)Online publication date: 21-Jan-2025
      • (2024)Beyond Recommendations: From Backward to Forward AI Support of Pilots' Decision-Making ProcessProceedings of the ACM on Human-Computer Interaction10.1145/36870248:CSCW2(1-32)Online publication date: 8-Nov-2024
      • (2024)Self-management of COPD supported by eHealth: Patients' attitudes towards monitoring, risk prediction and virtual coachingProceedings of the 13th Nordic Conference on Human-Computer Interaction10.1145/3679318.3685486(1-13)Online publication date: 13-Oct-2024
      • (2024)Understanding the User Perception and Experience of Interactive Algorithmic Recourse CustomizationACM Transactions on Computer-Human Interaction10.1145/367450331:3(1-25)Online publication date: 30-Aug-2024
      • (2024)VIME: Visual Interactive Model Explorer for Identifying Capabilities and Limitations of Machine Learning Models for Sequential Decision-MakingProceedings of the 37th Annual ACM Symposium on User Interface Software and Technology10.1145/3654777.3676323(1-21)Online publication date: 13-Oct-2024
      • (2024)Requirements and Attitudes towards Explainable AI in Law EnforcementProceedings of the 2024 ACM Designing Interactive Systems Conference10.1145/3643834.3661629(995-1009)Online publication date: 1-Jul-2024
      • Show More Cited By

      View Options

      View options

      PDF

      View or Download as a PDF file.

      PDF

      eReader

      View online with eReader.

      eReader

      HTML Format

      View this article in HTML Format.

      HTML Format

      Login options

      Figures

      Tables

      Media

      Share

      Share

      Share this Publication link

      Share on social media