ABSTRACT
Most explainable AI (XAI) approaches focus on mathematically correct and complete explanations. In contrast to that, everyday explanations are social practices that are influenced by many factors, e.g. the social and institutional context of the explanation. Everyday explanations are therefore co-constructed by both the explainer and the explainee. Assuming that everyday co-constructed explanations are easy to understand for citizens, they can help to make AI understandable for people who do not have affinity for technology. In this context, concrete and simplified examples that are adapted to the needs of the learner can reduce the complexity of the explanation furthermore.
This Ph.D. project investigates the effects of co-constructed explanations created while exploring notional machines on the mental models of AI of everyday people. The goal is to identify aspects of AI technology which are relevant for everyone, to empower people to become a responsible and active members of society in a digitalized world where AI technology is ubiqitious.
- Design-Based Research Collective. 2003. Design-based research: An emerging paradigm for educational inquiry. Educational researcher 32, 1 (2003), 5–8.Google Scholar
- Sally Fincher, Johan Jeuring, Craig S Miller, Peter Donaldson, Benedict Du Boulay, Matthias Hauswirth, Arto Hellas, Felienne Hermans, Colleen Lewis, Andreas Mühling, Janice L. Pearce, and Andrew Petersen. 2020. Notional Machines in Computing Education: The Education of Attention. In Proceedings of the Working Group Reports on Innovation and Technology in Computer Science Education. 21–50.Google ScholarDigital Library
- Urban Fraefel. 2014. Professionalization of pre-service teachers through univer-sity-school partnerships Partner schools for Professional Development»: development, imple-mentation and evaluation of cooperative learning in schools and classes. https://doi.org/10.13140/RG.2.1.1979.5925Google Scholar
- David Gunning, Mark Stefik, Jaesik Choi, Timothy Miller, Simone Stumpf, and Guang-Zhong Yang. 2019. XAI—Explainable artificial intelligence. Science Robotics 4, 37 (2019).Google Scholar
- Subbarao Kambhampati. 2021. Polanyi’s revenge and AI’s new romance with tacit knowledge. Commun. ACM 64, 2 (2021), 31–32.Google ScholarDigital Library
- Tim Miller. 2019. Explanation in artificial intelligence: Insights from the social sciences. Artificial intelligence 267 (2019), 1–38.Google Scholar
- Iyad Rahwan, Manuel Cebrian, Nick Obradovich, Josh Bongard, Jean-François Bonnefon, Cynthia Breazeal, Jacob W Crandall, Nicholas A Christakis, Iain D Couzin, Matthew O Jackson, Nicholas R. Jennings, Ece Kamar, Isabel M. Kloumann, Hugo Larochelle, David Lazer, Richard McElreath, Alan Mislove, David C. Parkes, Alex ’Sandy’ Pentland, Margaret E. Robert, Azim Shariff, Joshua B. Tenenbaum, and Michael Wellman. 2019. Machine behaviour. Nature 568, 7753 (2019), 477–486.Google Scholar
- Katharina J Rohlfing, Philipp Cimiano, Ingrid Scharlau, Tobias Matzner, Heike M Buhl, Hendrik Buschmeier, Elena Esposito, Angela Grimminger, Barbara Hammer, Reinhold Häb-Umbach, Ilona Horwath, Eyke Hüllermeier, Friederike Kern, Stefan Kopp, Kirsten Thommes, Axel-Cyrille Ngonga Ngomo, Carsten Schulte, Henning Wachsmuth, Petra Wagner, and Britta Wrede. 2020. Explanation as a social practice: Toward a conceptual framework for the social design of AI systems. IEEE Transactions on Cognitive and Developmental Systems (2020).Google Scholar
- Carsten Schulte and Lea Budde. 2018. A Framework for Computing Education: Hybrid Interaction System: The need for a bigger picture in computing education. In Proceedings of the 18th Koli Calling International Conference on Computing Education Research. 1–10.Google ScholarDigital Library
- Sue Sentance, Jane Waite, and Maria Kallia. 2019. Teaching computer programming with PRIMM: a sociocultural perspective. Computer Science Education 29, 2-3 (2019), 136–176.Google ScholarCross Ref
Recommendations
Artificial intelligence
AbstractArtificial intelligence (AI) is the Science and Engineering domain concerned with the theory and practice of developing systems that exhibit the characteristics we associate with intelligence in human behavior. Starting with a brief history of ...
Methods and standards for research on explainable artificial intelligence: Lessons from intelligent tutoring systems
AbstractThe DARPA Explainable Artificial Intelligence (AI) (XAI) Program focused on generating explanations for AI programs that use machine learning techniques. This article highlights progress during the DARPA Program (2017‐2021) relative to research ...
Lessons learned in the work on intelligent tutoring systems that apply to system design in Explainable AI. image image
Comments