Abstract
Medical image diagnosis system by using deep neural networks (DNN) can improve the sensitivity and speed of interpretation of chest CT for COVID-19 screening. However, DNN based medical image diagnosis is known to be influenced by the adversarial perturbations. In order to improve the robustness of medical image diagnosis system, this paper proposes an adversarial attack training method by using multi-loss hybrid adversarial function with heuristic projection. Firstly, the effective adversarial attacks which contain the noise style that can puzzle the network are created with a multi-loss hybrid adversarial function (MLAdv). Then, instead of adding these adversarial attacks to the training data directly, we consider the similarity between the original samples and adversarial attacks by using an adjacent loss during the training process, which can improve the robustness and the generalization of the network for unanticipated noise perturbations. Experiments are finished on COVID-19 dataset. The average attack success rate of this method for three DNN based medical image diagnosis systems is 63.9%, indicating that the created adversarial attack has strong attack transferability and can puzzle the network effectively. In addition, with the adversarial attack training, the augmented networks by using adversarial attacks can improve the diagnosis accuracy by 4.75%. Therefore, the augmented network based on MLAdv adversarial attacks can improve the robustness of medical image diagnosis system.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Paschali, M., Conjeti, S., Navarro, F., Navab, N.: Generalizability vs. robustness: investigating medical imaging networks using adversarial examples. In: Frangi, A.F., Schnabel, J.A., Davatzikos, C., Alberola-López, C., Fichtinger, G. (eds.) MICCAI 2018. LNCS, vol. 11070, pp. 493–501. Springer, Cham (2018). https://doi.org/10.1007/978-3-030-00928-1_56
Hein, M., Andriushchenko, M., Bitterwolf, J.: Why ReLU networks yield high-confidence predictions far away from the training data and how to mitigate the problem. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 41–50 (2019)
Goodfellow, I.J., Shlens, J., Szegedy, C.: Explaining and harnessing adversarial examples. In: ICLR (2015)
Kurakin, A., Goodfellow, I.J., Bengio, S.: Adversarial machine learning at scale. CoRR: abs/1611.01236 (2016)
Tramèr, F., Kurakin, A., Papernot, N., Goodfellow, I., Boneh, D., McDaniel, P.: Ensemble adversarial training: attacks and defenses. arXiv:1705.07204 (2017)
Akhtar, N., Liu, J., Mian, A.: Defense against universal adversarial perturbations. In: CVPR, pp. 3389–3398 (2018)
Sarki, R., Ahmed, K.: Wang, H: Image preprocessing in classification and identification of diabetic eye diseases. Data Sci. Eng. 6(4), 455–471 (2021)
Papernot, N., McDaniel, P., Wu, X., Jha, S., Swami, A.: Distillation as a defense to adversarial perturbations against deep neural networks. In: IEEE Symposium on Security and Privacy, pp. 582–597 (2016)
Sarki, R., Ahmed, K., Wang, H., Zhang, Y., Ma, J., Wang, K.: Image preprocessing in classification and identification of diabetic eye diseases. Data Sci. Eng. 6(4), 455–471 (2021). https://doi.org/10.1007/s41019-021-00167-z
Sharif, M., Bhagavatula, S., Bauer, L., et al.: Accessorize to a crime: real and stealthy attacks on state-of-the-art face recognition. In: Proceedings of the 2016 ACM SIGSAC Conference on Computer and Communications Security, pp. 1528–1540 (2016)
Selvaraju, R.R., Cogswell, M., Das, A., et al.: Grad-CAM: visual explanations from deep networks via gradient-based localization. In: Proceedings of the IEEE international conference on computer vision, pp. 618–626 (2017)
Mahendran, A., Vedaldi A.: Understanding deep image representations by inverting them. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 5188–5196 (2015)
Li, Y., Bai, S., Xie, C., Liao, Z., Shen, X., Yuille, A.: Regional homogeneity: towards learning transferable universal adversarial perturbations against defenses. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12356, pp. 795–813. Springer, Cham (2020). https://doi.org/10.1007/978-3-030-58621-8_46
Rosen, J.B.: The gradient projection method for nonlinear programming. Part I. Linear constraints. J. Soc. Ind. Appl. Math. 8(1), 181–217 (1960)
https://github.com/ml-workgroup/covid-19-image-repository/tree/master/png
https://www.kaggle.com/paultimothymooney/chest-xray-pneumonia
Acknowledgement
This work is supported in part by the National Nature Science Foundation of China grants (U20A20389, 61901214).
Author information
Authors and Affiliations
Corresponding authors
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2023 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this paper
Cite this paper
Cheng, C., Chen, F. (2023). Improving Robustness of Medical Image Diagnosis System by Using Multi-loss Hybrid Adversarial Function with Heuristic Projection. In: Li, B., Yue, L., Tao, C., Han, X., Calvanese, D., Amagasa, T. (eds) Web and Big Data. APWeb-WAIM 2022. Lecture Notes in Computer Science, vol 13422. Springer, Cham. https://doi.org/10.1007/978-3-031-25198-6_31
Download citation
DOI: https://doi.org/10.1007/978-3-031-25198-6_31
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-25197-9
Online ISBN: 978-3-031-25198-6
eBook Packages: Computer ScienceComputer Science (R0)