Abstract
Convolutional neural networks (CNNs) may learn spurious correlations between bias features (e.g., background) and labels in image classification. The spuriousness in CNNs usually occurs in building connections between the background of images and labels. Such spurious correlation limits the generalizability of CNNs in classification tasks. Changing backgrounds and foregrounds of original samples can reduce the spuriousness in natural image classification. However, generating annotations for foreground on medical image datasets is time-consuming and labor-intensive. To solve this problem, we propose an online data augmentation method named Combining Foreground And Background (CFAB), which makes CNNs focus on key causal features without foreground annotations and breaks the correlation between backgrounds and labels by changing different backgrounds for one sample. Furthermore, we propose a framework for collaborative augmenting samples using CFAB and training CNNs. Comprehensive experiments indicate that the proposed method weakens the spuriousness, improves the generalizability of the model, and achieves state-of-the-art results in medical ultrasound dataset classification.
The authors acknowledge that this work was supported in part by the National Natural Science Foundation of China (NSFC) No. 82071930.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
References
Al-Dhabyani, W., Gomaa, M., Khaled, H., Fahmy, A.: Dataset of breast ultrasound images. Data Brief 28, 104863 (2020)
Chang, C.H., Adam, G.A., Goldenberg, A.: Towards robust classification model by counterfactual and invariant data generation. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 15212–15221 (2021)
Chen, Y., Bai, Y., Zhang, W., Mei, T.: Destruction and construction learning for fine-grained image recognition. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 5157–5166 (2019)
Geirhos, R., et al.: Shortcut learning in deep neural networks. Nat. Mach. Intell. 2(11), 665–673 (2020)
Grigorescu, S., Trasnea, B., Cocias, T., Macesanu, G.: A survey of deep learning techniques for autonomous driving. J. Field Rob. 37(3), 362–386 (2020)
Jiang, P.T., Zhang, C.B., Hou, Q., Cheng, M.M., Wei, Y.: Layercam: exploring hierarchical class activation maps for localization. IEEE Trans. Image Process. 30, 5875–5888 (2021)
Kim, D., Cho, D., Yoo, D., So Kweon, I.: Two-phase learning for weakly supervised object localization. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 3534–3543 (2017)
Kim, J.H., Choo, W., Song, H.O.: Puzzle mix: exploiting saliency and local statistics for optimal mixup. In: International Conference on Machine Learning, pp. 5275–5285. PMLR (2020)
Kim, J., Choo, W., Jeong, H., Song, H.O.: Co-mixup: saliency guided joint mixup with supermodular diversity. In: International Conference on Learning Representations (2021). https://openreview.net/forum?id=gvxJzw8kW4b
Li, X., Liang, H., Nagala, S., Chen, J.: Improving ultrasound image classification with local texture quantisation. In: IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 1211–1215. IEEE (2022)
Moreno-Barea, F.J., Jerez, J.M., Franco, L.: Improving classification accuracy using data augmentation on small data sets. Expert Syst. Appl. 161, 113696 (2020)
Nalepa, J., Myller, M., Kawulok, M.: Hyperspectral data augmentation. arXiv preprint arXiv:1903.05580 (2019)
Qin, J., Fang, J., Zhang, Q., Liu, W., Wang, X., Wang, X.: Resizemix: mixing data with preserved object information and true labels. arXiv: Computer Vision and Pattern Recognition (2020)
Radiology, A., D’Orsi, C.: ACR BI-RADS Atlas: breast imaging reporting and data system; mammography, ultrasound, magnetic resonance imaging, follow-up and outcome monitoring, data dictionary. ACR, American College of Radiology (2013)
Selvaraju, R.R., Cogswell, M., Das, A., Vedantam, R., Parikh, D., Batra, D.: Grad-cam: visual explanations from deep networks via gradient-based localization. In: Proceedings of the IEEE International Conference on Computer Vision (ICCV) (2017)
Shen, D., Wu, G., Suk, H.I.: Deep learning in medical image analysis. Ann. Review Biomed. Eng. 19, 221–248 (2017)
Yun, S., Han, D., Oh, S.J., Chun, S., Choe, J., Yoo, Y.: Cutmix: regularization strategy to train strong classifiers with localizable features. In: Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 6023–6032 (2019)
Zhang, H., Cisse, M., Dauphin, Y.N., Lopez-Paz, D.: mixup: beyond empirical risk minimization. In: International Conference on Learning Representations (2018). https://openreview.net/forum?id=r1Ddp1-Rb
Zhou, B., Khosla, A., Lapedriza, A., Oliva, A., Torralba, A.: Learning deep features for discriminative localization. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 2921–2929 (2016)
Author information
Authors and Affiliations
Corresponding authors
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2023 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this paper
Cite this paper
Huang, J., Huang, K., Xu, M., Liu, F. (2023). CFAB: An Online Data Augmentation to Alleviate the Spuriousness of Classification on Medical Ultrasound Images. In: Christensen, H.I., Corke, P., Detry, R., Weibel, JB., Vincze, M. (eds) Computer Vision Systems. ICVS 2023. Lecture Notes in Computer Science, vol 14253. Springer, Cham. https://doi.org/10.1007/978-3-031-44137-0_8
Download citation
DOI: https://doi.org/10.1007/978-3-031-44137-0_8
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-44136-3
Online ISBN: 978-3-031-44137-0
eBook Packages: Computer ScienceComputer Science (R0)