Abstract
Despite their remarkable revolution in artificial intelligence and promising performance, deep neural networks are susceptible to biases due to imbalanced training data or the way that training data is represented as features for the model; thus, they need a high level of transparency in critical fields, especially in the medical sector. Several explainability methods have been developed, aiming at enhancing the transparency of the model and increasing its trustworthiness. One of these methods is Grad-CAM, which can highlight the significant parts of an input that most influence the model’s prediction and detect biases in the dataset. In this paper, a method is provided for integrating human knowledge with the deep neural network to improve its performance, reduce the biases that arise, and leverage human experience. This can be done via the attention maps generated by Grad-CAM, which are then edited by domain experts. The proposed method enables the edited attention maps to contribute to the training of the model and updating its parameters, leading to better generalisation and fewer inductive biases. This approach is applied to several image classification datasets: Imagenette2, BUSI, and skin cancer ISIC. Experimental results show that the generated attention maps highly correspond to the edited ones while having high classification performance.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Al-Dhabyani, W., Gomaa, M., Khaled, H., Fahmy, A.: Dataset of breast ultrasound images. Data Brief 28, 104863 (2020)
Karar, A., Shaikh, Z.A., Khan, A.A., Laghari, A.A.: Multiclass skin cancer classification using EfficientNets-a first step towards preventing skin cancer. Neurosci. Inform. 2(4), 100034 (2022)
Fukui, H., Hirakawa, T., Yamashita, T., Fujiyoshi, H.: Attention branch network: learning of attention mechanism for visual explanation. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 10705–10714 (2019)
Howard, J.: Imagenette: a smaller subset of 10 easily classified classes from ImageNet, and a little more French (2019)
Dong, W., Deng, J., Li, L.-J., Socher, R., Li, K., Fei-Fei, L.: A large-scale hierarchical image database. In: IEEE Conference on Computer Vision and Pattern Recognition, pp. 248–255 (2009)
Keswani, V., Lease, M., Kenthapadi, K.: Designing closed human-in-the-loop deferral pipelines. arXiv preprint arXiv:2202.04718 (2022)
Lin, M., Chen, Q., Yan, S.: Network in network. arXiv preprint arXiv:1312.4400 (2013)
Mitsuhara, M., et al.: Embedding human knowledge into deep neural network via attention map. arXiv preprint arXiv:1905.03540 (2019)
Monarch, R.M.: Human-in-the-Loop Machine Learning: Active learning and annotation for human-centered AI. Simon and Schuster (2021)
Montavon, Grégoire., Binder, Alexander, Lapuschkin, Sebastian, Samek, Wojciech, Müller, Klaus-Robert.: Layer-wise relevance propagation: an overview. In: Samek, Wojciech, Montavon, Grégoire., Vedaldi, Andrea, Hansen, Lars Kai, Müller, Klaus-Robert. (eds.) Explainable AI: Interpreting, Explaining and Visualizing Deep Learning. LNCS (LNAI), vol. 11700, pp. 193–209. Springer, Cham (2019). https://doi.org/10.1007/978-3-030-28954-6_10
Ras, G., Xie, N., Van Gerven, M., Doran, D.: Explainable deep learning: a field guide for the uninitiated. J. Artif. Intell. Res. 73, 329–397 (2022)
Selvaraju, R.R., Cogswell, M., Das, A., Vedantam, R., Parikh, D., Batra, D.: Grad-CAM: visual explanations from deep networks via gradient-based localization. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 618–626 (2017)
Springenberg, J.T., Dosovitskiy, A., Brox, T., Riedmiller, M.: Striving for simplicity: the all convolutional net. arXiv preprint arXiv:1412.6806 (2014)
Tsiakas, K., Murray-Rust, D.: Using human-in-the-loop and explainable AI to envisage new future work practices. In: Proceedings of the 15th International Conference on Pervasive Technologies Related to Assistive Environments, pp. 588–594 (2022)
Zhou, B., Khosla, A., Lapedriza, A., Oliva, A., Torralba, A.: Learning deep features for discriminative localization. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 2921–2929 (2016)
Author information
Authors and Affiliations
Corresponding authors
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2023 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this paper
Cite this paper
Elbeialy, H. et al. (2023). Visual Steering for Deep Neural Networks Using Explainable Artificial Intelligence. In: Hassanien, A., Rizk, R.Y., Pamucar, D., Darwish, A., Chang, KC. (eds) Proceedings of the 9th International Conference on Advanced Intelligent Systems and Informatics 2023. AISI 2023. Lecture Notes on Data Engineering and Communications Technologies, vol 184. Springer, Cham. https://doi.org/10.1007/978-3-031-43247-7_4
Download citation
DOI: https://doi.org/10.1007/978-3-031-43247-7_4
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-43246-0
Online ISBN: 978-3-031-43247-7
eBook Packages: Intelligent Technologies and RoboticsIntelligent Technologies and Robotics (R0)