Abstract
With the development of medical imaging technologies, breast cancer segmentation remains challenging, especially when considering multimodal imaging. Compared to a single-modality image, multimodal data provide additional information, contributing to better representation learning capabilities. This paper applies these advantages by presenting a deep learning network architecture for segmenting breast cancer with multimodal computed tomography (CT) images based on fusing U-Net architectures that can learn richer representations from multimodal data. The multipath fusion architecture introduces an additional fusion module across different paths, enabling the model to extract features from different modalities at each level of the encoding path. This approach enhances segmentation performance and produces more robust results compared to using a single modality. The study reports experiments conducted on multimodal CT images from 36 patients for training, validation, and testing purposes. The results demonstrate that the proposed model ouperforms the U-Net architecture when considering different combinations of input image modalities. Specifically, when combining two distinct CT modalities, the ZE and IoNW input combination yields the highest Dice score of 0.8546.
Similar content being viewed by others
Data availability
All requests for raw and analyzed data will be made available upon reasonable request for academic use and within the limitations of the provided informed consent by the corresponding author upon acceptance. Every request will be reviewed by the institutional review board of the School of Electrical Engineering of Southwest Jiaotong University and the affiliated hospital of Southwest Medical University.
Notes
Github code will be made available for reference at https://github.com/AisenCD/dualCT.
References
Bray F, Ferlay J, Soerjomataram I, Siegel RL, Torre LA, Jemal A (2018) Global cancer statistics 2018: Globocan estimates of incidence and mortality worldwide for 36 cancers in 185 countries. CA a cancer journal for clinicians 68(6):394–424
Zou L, Yu S, Meng T, Zhang Z, Liang X, Xie Y (2019) A technical review of convolutional neural network-based mammographic breast cancer diagnosis. Comput Math Methods Med 2019
Ye C, Wang W, Zhang S, Wang K (2019) Multi-depth fusion network for whole-heart ct image segmentation. IEEE Access 7:23421–23429
Hua C-h, Shapira N, Merchant TE, Klahr P, Yagil Y (2018) Accuracy of electron density, effective atomic number, and iodine concentration determination with a dual-layer dual-energy computed tomography system. Med Phys 45(6):2486–2497
Michael E, Ma H, Li H, Kulwa F, Li J (2021) Breast cancer segmentation methods: current status and future potentials. BioMed Res Int 2021
Bai J, Posner R, Wang T, Yang C, Nabavi S (2021) Applying deep learning in digital breast tomosynthesis for automatic breast cancer detection: a review. Med Image Anal 71:102049
Ronneberger O, Fischer P, Brox T (2015) U-net: convolutional networks for biomedical image segmentation. In: International conference on medical image computing and computer-assisted intervention, Springer, pp 234–241
Salama WM, Aly MH (2021) Deep learning in mammography images segmentation and classification: automated cnn approach. Alex Eng J 60(5):4701–4709
Ravitha Rajalakshmi N, Vidhyapriya R, Elango N, Ramesh N (2021) Deeply supervised u-net for mass segmentation in digital mammograms. Int J Imaging Syst Technol 31(1):59–71
Abdelhafiz D, Bi J, Ammar R, Yang C, Nabavi S (2020) Convolutional neural network for automated mass segmentation in mammography. BMC bioinformatics 21(1):1–19
Hossain MS (2019) Microc alcification segmentation using modified u-net segmentation network from mammogram images. J King Saud Univ-Comput Inform Sci
Li J, Dong D, Fang M, Wang R, Tian J, Li H, Gao J (2020) Dual-energy ct-based deep learning radiomics can improve lymph node metastasis risk prediction for gastric cancer. Eur Radiol 30(4):2324–2333
An C, Li D, Li S, Li W, Tong T, Liu L, Jiang D, Jiang L, Ruan G, Hai N et al (2022) Deep learning radiomics of dual-energy computed tomography for predicting lymph node metastases of pancreatic ductal adenocarcinoma. Eur J Nucl Med Mol Imaging 49(4):1187–1199
Wang Y-W, Chen C-J, Wang T-C, Huang H-C, Chen H-M, Shih J-Y, Chen J-S, Huang Y-S, Chang Y-C, Chang R-F (2022) Multi-energy level fusion for nodal metastasis classification of primary lung tumor on dual energy ct using deep learning. Comput Biol Med 141:105185
Zhang W, Li R, Deng H, Wang L, Lin W, Ji S, Shen D (2015) Deep convolutional neural networks for multi-modality isointense infant brain image segmentation. NeuroImage 108:214–224
Dolz J, Ayed IB, Yuan J, Desrosiers C (2018) Isointense infant brain segmentation with a hyper-dense connected convolutional neural network. In: 2018 IEEE 15th International symposium on biomedical imaging (ISBI 2018), IEEE, pp 616–620
Nie D, Wang L, Gao Y, Shen D (2016) Fully convolutional networks for multi-modality isointense infant brain image segmentation. In: 2016 IEEE 13Th International symposium on biomedical imaging (ISBI), IEEE, pp 1342–1345
Dolz J, Desrosiers C, Ben Ayed I (2018) Ivd-net: intervertebral disc localization and segmentation in mri with a multi-modal unet. In: International workshop and challenge on computational methods and clinical applications for Spine imaging, Springer, pp 130–143
Große Hokamp N, Lennartz S, Salem J, Pinto dos Santos D, Heidenreich A, Maintz D, Haneder S (2020) Dose independent characterization of renal stones by means of dual energy computed tomography and machine learning: an ex-vivo study. Eur Radiol 30:1397–1404
Choi B, Choi IY, Yeom Cha SH, SK, Chung HH, Lee SH, Cha J, Lee J-H (2020) Feasibility of computed tomography texture analysis of hepatic fibrosis using dual-energy spectral detector computed tomography. Japan J Radiology 38:1179–1189
Shapira N, Fokuhl J, Schultheiß M, Beck S, Kopp FK, Pfeiffer D, Dangelmaier J, Pahn G, Sauter AP, Renger B et al (2020) Liver lesion localisation and classification with convolutional neural networks: a comparison between conventional and spectral computed tomography. Biomedical Physics & Engineering Express 6(1):015038
Alom MZ, Taha TM, Yakopcic C, Westberg S, Sidike P, Nasrin MS, Van Esesn BC, Awwal AAS, Asari VK (2018) The history began from alexnet: a comprehensive survey on deep learning approaches. arXiv:1803.01164
He K, Zhang X, Ren S, Sun J (2016) Deep residual learning for image recognition. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 770–778
Milletari F, Navab N, Ahmadi S-A (2016) V-net: fully convolutional neural networks for volumetric medical image segmentation. In: 2016 Fourth international conference on 3d vision (3DV), IEEE, pp 565–571
Han X (2017) Automatic liver lesion segmentation using a deep convolutional neural network method. arXiv:1704.07239
Yu S, Xiao D, Frost S, Kanagasingam Y (2019) Robust optic disc and cup segmentation with deep learning for glaucoma detection. Comput Med Imaging Grap 74:61–71
Kong Z, Xiong F, Zhang C, Fu Z, Zhang M, Weng J, Fan M (2020) Automated maxillofacial segmentation in panoramic dental x-ray images using an efficient encoder-decoder network. IEEE Access 8:207822–207833
Pham V-T, Tran T-T, Wang P-C, Chen P-Y, Lo M-T (2021) Ear-unet: a deep learning-based approach for segmentation of tympanic membranes from otoscopic images. Artif Intell Med 115:102065
Gao S-H, Cheng M-M, Zhao K, Zhang X-Y, Yang M-H, Torr P (2019) Res2net: a new multi-scale backbone architecture. IEEE Trans Pattern Anal Mach Intell 43(2):652–662
Qin J, Huang Y, Wen W (2020) Multi-scale feature fusion residual network for single image super-resolution. Neurocomputing 379:334–342
Fu J, Li W, Du J, Huang Y (2021) A multiscale residual pyramid attention network for medical image fusion. Biomed Signal Process Control 66:102488
Ma Y, Qi F, Wang P, Liang F, Lv H, Yu X, Li Z, Xue H, Wang J, Zhang Y (2020) Multiscale residual attention network for distinguishing stationary humans and common animals under through-wall condition using ultra-wideband radar. IEEE Access 8:121572–121583
Zhang H, Wu C, Zhang Z, Zhu Y, Lin H, Zhang Z, Sun Y, He T, Mueller J, Manmatha R et al (2020) Resnest: split-attention networks. arXiv:2004.08955
Zhang J, Zhang J, Hu G, Chen Y, Yu S (2019) Scalenet: a convolutional network to extract multi-scale and fine-grained visual features. IEEE Access 7:147560–147570
Liu H, Cao H, Song E, Ma G, Xu X, Jin R, Jin Y, Hung C-C (2019) A cascaded dual-pathway residual network for lung nodule segmentation in ct images. Physica Medica 63:112–121
Ioffe S, Szegedy C (2015) Batch normalization: accelerating deep network training by reducing internal covariate shift. In: International conference on machine learning, pp 448–456 PMLR
He K, Zhang X, Ren S, Sun J (2015) Delving deep into rectifiers: Surpassing human-level performance on imagenet classification. In: Proceedings of the IEEE international conference on computer vision, pp 1026–1034
van Ommen F, Bennink E, Vlassenbroek A, Dankbaar JW, Schilham AM, Viergever MA, de Jong HW (2018) Image quality of conventional images of dual-layer spectral ct: a phantom study. Med Phys 45(7):3031–3042
Ghaffari M, Sowmya A, Oliver R (2019) Automated brain tumor segmentation using multimodal brain scans: a survey based on models submitted to the brats 2012–2018 challenges. IEEE Rev Biomed Eng 13:156–168
Taha AA, Hanbury A (2015) An efficient algorithm for calculating the exact hausdorff distance. IEEE Trans Pattern Anal Mach Intell 37(11):2153–2163
Koonce B, Koonce B (2021) Mobilenetv3. Image recognition and dataset categorization, convolutional neural networks with Swift for Tensorflow, pp 125–144
Ibtehaz N, Rahman MS (2020) Multiresunet: rethinking the u-net architecture for multimodal biomedical image segmentation. Neural Netw 121:74–87
Szegedy C, Vanhoucke V, Ioffe S, Shlens J, Wojna Z (2016) Rethinking the inception architecture for computer vision. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 2818–2826
Acknowledgements
This work was financially aided by the Central Universities Pay for Basic Scientific Research (2682021ZTPY027), Natural Science Foundation of China (62173279, U1934221) and Sichuan Science and Technology Program under Grant 2022YFG0247, 2021JDJQ0012, 2020YFQ0057.
Author information
Authors and Affiliations
Corresponding authors
Ethics declarations
Conflicts of interests
All authors declare that they have no conflicts of interest.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
About this article
Cite this article
Yang, A., Xu, L., Qin, N. et al. MFU-Net: a deep multimodal fusion network for breast cancer segmentation with dual-layer spectral detector CT. Appl Intell 54, 3808–3824 (2024). https://doi.org/10.1007/s10489-023-05090-6
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10489-023-05090-6