Abstract
Benefit from the progress in the field of explainable artificial intelligence (XAI), explanations have been increasingly prospective in the autonomous vehicle (AV) context. Providing explanations has been proved to be vital for human-AV interaction, but what and how to explain are still to be addressed. This study seeks to bridge the areas of XAI and human-AV interaction by combining perspectives of both users and researchers. In this paper, a conceptual framework of explanation models was proposed to indicate what aspects to explain in human-AV interaction. Based on the framework, we introduced a scenario-based and question-driven method, i.e., the SQX-canvas, to guide the workflow of generating explanations from users’ demands in a certain AV scenario. To make an initial validation of the method, a co-design workshop involving researchers and users was conducted with four AV scenarios provided in forms of video clips. Participants produced explanation concepts and expressed their attitudes towards the AV scenarios following the “scenario, question and explanation” process. It was apparent that users’ demands of explanations varied across scenarios, and findings as well as limitations were discussed. This method could provide implications for research and practice on facilitating transparent human-AV interaction.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Adadi, A., Berrada, M.: Peeking inside the black-box: a survey on explainable artificial intelligence (XAI). IEEE Access 6, 52138–52160 (2018)
Arya, V., et al.: One explanation does not fit all: a toolkit and taxonomy of AI explainability techniques. arXiv preprint arXiv:1909.03012 (2019)
Atakishiyev, S., Salameh, M., Yao, H., Goebel, R.: Explainable artificial intelligence for autonomous driving: a comprehensive overview and field guide for future research directions. arXiv preprint arXiv:2112.11561 (2021)
Banks, V.A., Plant, K.L., Stanton, N.A.: Driver error or designer error: using the Perceptual Cycle Model to explore the circumstances surrounding the fatal Tesla crash on 7th May 2016. Saf. Sci. 108, 278–285 (2018). https://doi.org/10.1016/j.ssci.2017.12.023. https://www.sciencedirect.com/science/article/pii/S0925753517314212
Carsten, O., Martens, M.H.: How can humans understand their automated cars? HMI principles, problems and solutions. Cognit. Technol. Work 21(1), 3–20 (2019). https://doi.org/10.1007/s10111-018-0484-0
De Winter, J.C., Happee, R., Martens, M.H., Stanton, N.A.: Effects of adaptive cruise control and highly automated driving on workload and situation awareness: a review of the empirical evidence. Transport. Res. F: Traffic Psychol. Behav. 27, 196–217 (2014)
Debernard, S., Chauvin, C., Pokam, R., Langlois, S.: Designing human-machine interface for autonomous vehicles. IFAC-PapersOnLine 49(19), 609–614 (2016)
Du, M., Liu, N., Hu, X.: Techniques for interpretable machine learning. Commun. ACM 63(1), 68–77 (2019)
Du, N., et al.: Look who’s talking now: implications of AV’S explanations on driver’s trust, AV preference, anxiety and mental workload. Transp. Res. Part C: Emerg. Technol. 104, 428–442 (2019)
Eiband, M., Schneider, H., Bilandzic, M., Fazekas-Con, J., Haug, M., Hussmann, H.: Bringing transparency design into practice. In: 23rd International Conference on Intelligent User Interfaces, pp. 211–223 (2018)
Eriksson, A., Petermeijer, S.M., Zimmermann, M., De Winter, J.C., Bengler, K.J., Stanton, N.A.: Rolling out the red (and green) carpet: supporting driver decision making in automation-to-manual transitions. IEEE Trans. Human-Mach. Syst. 49(1), 20–31 (2018)
Geyer, S., et al.: Concept and development of a unified ontology for generating test and use-case catalogues for assisted and automated vehicle guidance. IET Intel. Transp. Syst. 8(3), 183–189 (2014)
SAE International: J3016C: Taxonomy and Definitions for Terms Related to Driving Automation Systems for On-Road Motor Vehicles - SAE International (2021). https://www.sae.org/standards/content/j3016202104/
Kircher, K., Larsson, A., Hultgren, J.A.: Tactical driving behavior with different levels of automation. IEEE Trans. Intell. Transp. Syst. 15(1), 158–167 (2014). https://doi.org/10.1109/TITS.2013.2277725. Conference Name: IEEE Transactions on Intelligent Transportation Systems
Koo, J., Kwac, J., Ju, W., Steinert, M., Leifer, L., Nass, C.: Why did my car just do that? Explaining semi-autonomous driving actions to improve driver understanding, trust, and performance. Int. J. Interact. Design Manuf. (IJIDeM) 9(4), 269–275 (2015)
Liao, Q.V., Gruen, D., Miller, S.: Questioning the AI: informing design practices for explainable AI user experiences. In: Proceedings of the 2020 CHI Conference on Human Factors in Computing Systems, pp. 1–15 (2020)
Lyons, J.B.: Being transparent about transparency: a model for human-robot interaction. In: 2013 AAAI Spring Symposium Series (2013)
Merat, N., et al.: The “out-of-the-loop’’ concept in automated driving: proposed definition, measures and implications. Cogn. Technol. Work 21(1), 87–98 (2019). https://doi.org/10.1007/s10111-018-0525-8
Miller, T.: Explanation in artificial intelligence: insights from the social sciences. Artif. Intell. 267, 1–38 (2019)
Naujoks, F., Wiedemann, K., Schömig, N., Hergeth, S., Keinath, A.: Towards guidelines and verification methods for automated vehicle HMIS. Transport. Res. F: Traffic Psychol. Behav. 60, 121–136 (2019)
Omeiza, D., Webb, H., Jirotka, M., Kunze, L.: Explanations in autonomous driving: a survey. IEEE Trans. Intell. Transp. Syst. (2021)
Pokam, R., Debernard, S., Chauvin, C., Langlois, S.: Principles of transparency for autonomous vehicles: first results of an experiment with an augmented reality human-machine interface. Cogn. Technol. Work 21(4), 643–656 (2019). https://doi.org/10.1007/s10111-019-00552-9
Rezvani, T., Driggs-Campbell, K., Bajcsy, R.: Optimizing interaction between humans and autonomy via information constraints on interface design. In: 2017 IEEE 20th International Conference on Intelligent Transportation Systems (ITSC), pp. 1–6 (2017). https://doi.org/10.1109/ITSC.2017.8317686. iSSN: 2153-0017
Schneider, T., Hois, J., Rosenstein, A., Ghellal, S., Theofanou-Fülbier, D., Gerlicher, A.R.: Explain yourself! transparency for positive UX in autonomous driving. In: Proceedings of the 2021 CHI Conference on Human Factors in Computing Systems, pp. 1–12 (2021)
Schömig, N., et al.: Checklist for expert evaluation of HMIS of automated vehicles-discussions on its value and adaptions of the method within an expert workshop. Information 11(4), 233 (2020)
Shen, Y., et al.: To explain or not to explain: a study on the necessity of explanations for autonomous vehicles. arXiv preprint arXiv:2006.11684 (2020)
Ulbrich, S., Menzel, T., Reschka, A., Schuldt, F., Maurer, M.: Defining and substantiating the terms scene, situation, and scenario for automated driving. In: 2015 IEEE 18th International Conference on Intelligent Transportation Systems, pp. 982–988. IEEE (2015)
Wang, D., Yang, Q., Abdul, A., Lim, B.Y.: Designing theory-driven user-centric explainable AI. In: Proceedings of the 2019 CHI Conference on Human Factors in Computing Systems, pp. 1–15 (2019)
Wiegand, G., Schmidmaier, M., Weber, T., Liu, Y., Hussmann, H.: I drive-you trust: explaining driving behavior of autonomous cars. In: Extended Abstracts of the 2019 CHI Conference on Human Factors in Computing Systems, pp. 1–6 (2019)
Wolf, C.T.: Explainability scenarios: towards scenario-based XAI design. In: Proceedings of the 24th International Conference on Intelligent User Interfaces, pp. 252–257 (2019)
Yan, F., Karaosmanoglu, S., Demir, A., Baumann, M.: Spatial visualization of sensor information for automated vehicles. In: Proceedings of the 11th International Conference on Automotive User Interfaces and Interactive Vehicular Applications: Adjunct Proceedings, pp. 265–270 (2019)
Yeong, D.J., Velasco-Hernandez, G., Barry, J., Walsh, J., et al.: Sensor and sensor fusion technology in autonomous vehicles: a review. Sensors 21(6), 2140 (2021)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2022 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this paper
Cite this paper
Zhang, Y., Guo, W., Chi, C., Hou, L., Sun, X. (2022). Towards Scenario-Based and Question-Driven Explanations in Autonomous Vehicles. In: Krömker, H. (eds) HCI in Mobility, Transport, and Automotive Systems. HCII 2022. Lecture Notes in Computer Science, vol 13335. Springer, Cham. https://doi.org/10.1007/978-3-031-04987-3_7
Download citation
DOI: https://doi.org/10.1007/978-3-031-04987-3_7
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-04986-6
Online ISBN: 978-3-031-04987-3
eBook Packages: Computer ScienceComputer Science (R0)