Abstract
Multimodal medical image fusion (MMIF) integrates the advantages of multiple source images to assist clinical diagnosis. Existing image fusion methods need help to distinguish the importance between features and often define features to be retained subjectively, which leads to global structure loss and limits the performance of fusion. To overcome these restrictions, we propose a novel self-supervised tensor low-rank decomposition fusion network that can effectively extract global information from high-rank to low-rank conversion processes. Specifically, the compensation of textural features is performed by employing a self-supervised auxiliary task, and the whole network is dynamically fine-tuned according to a hybrid loss. In our model, an enhanced weights (EW) estimation method based on the global luminance contrast is developed, and a structure tensor loss with constraints is introduced to improve the robustness of the fusion results. Moreover, extensive experiments on six types of multimodal medical images show that visual and qualitative results are superior to competitors, validating the effectiveness of our methods.









Similar content being viewed by others
Explore related subjects
Discover the latest articles, news and stories from top researchers in related subjects.References
Ding AS, Lu A, Li Z, Galaiya D, Siewerdsen JH, Taylor RH, Creighton FX (2022) Automated registration-based temporal bone computed tomography segmentation for applications in neurotologic surgery. Otolaryngol Head Neck Surg 167(1):133–140
Wu HT, Zheng K, Huang Q, Hu J (2021) Contrast enhancement of multiple tissues in mr brain images with reversibility. IEEE Signal Process Lett 28:160–164. https://doi.org/10.1109/LSP.2020.3048840
Valladares A, Beyer T, Rausch I (2020) Physical imaging phantoms for simulation of tumor heterogeneity in pet, ct, and mri: an overview of existing designs. Med Phys 47(4):2023–2037
Zhang M, Chu C, Huang L, Hu B (2022) Ct-mr image fusion for post-implant dosimetry analysis in brain tumor seed implantation-a preliminary study. Dis Markers 2022
Liu X, Li W, Liu Z, Du F, Zou Q (2021) A dual-branch model for diagnosis of parkin-son’s disease based on the independent and joint features of the left and right gait. Appl Intell 1–12
Říha P, Doležalová I, Mareček R, Lamoš M, Bartoňová M, Kojan M, Mikl M, Gajdoš M, Vojtíšek L, Bartoň M et al (2022) Multimodal combination of neuroimaging methods for localizing the epileptogenic zone in mr-negative epilepsy. Sci Rep 12(1):15158
Lu F, Du L, Chen W, Jiang H, Yang C, Pu Y, Wu J, Zhu J, Chen T, Zhang X et al (2022) T 1-t 2 dual-modal magnetic resonance contrast-enhanced imaging for rat liver fibrosis stage. RSC Adv 12(55):35809–35819
Ullah H, Zhao Y, Abdalla FY, Wu L (2022) Fast local laplacian filtering based enhanced medical image fusion using parameter-adaptive pcnn and local features-based fuzzy weighted matrices. Appl Intell 1–20
Yin M, Liu X, Liu Y, Chen X (2018) Medical image fusion with parameter-adaptive pulse coupled neural network in nonsubsampled shearlet transform domain. IEEE Trans Instrum Meas 68(1):49–64
Vanitha K, Satyanarayana D, Prasad MG (2021) Multi-modal medical image fusion algorithm based on spatial frequency motivated pa-pcnn in the nsst domain. Curr Med Imaging Rev 17(5):634–643
Nie R, Cao J, Zhou D, Qian W (2020) Multi-source information exchange encoding with pcnn for medical image fusion. IEEE Trans Circuits Syst Video Technol 31(3):986–1000
Chen L, Wang X, Zhu Y, Nie R (2022) Multi-level difference information replenishment for medical image fusion. Appl Intell 1–13
Zhao Z, Bai H, Zhang J, Zhang Y, Xu S, Lin Z, Timofte R, Van Gool L (2023) Cddfuse: Correlation-driven dual-branch feature decomposition for multi-modality image fusion. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 5906–5916
Ding Z, Li H, Guo Y, Zhou D, Liu Y, Xie S (2023) M4fnet: Multimodal medical image fusion network via multi-receptive-field and multi-scale feature integration. Comput Biol Med 159:106923
Zhang G, Nie R, Cao J (2022) Ssl-waeie: Self-supervised learning with weighted auto-encoding and information exchange for infrared and visible image fusion. EEE/CAA J Autom Sin 9(9):1694–1697
Zhang Y, Nie R, Cao J, Ma C, Wang C (2023) Ss-ssan: a self-supervised subspace attentional network for multi-modal medical image fusion. Artif Intell Rev 1–23
Liang P, Jiang J, Liu X, Ma J (2022) Fusion from decomposition: A self-supervised decomposition approach for image fusion. In: European Conference on Computer Vision, pp. 719–735 Springer
Aghamaleki JA, Ghorbani A (2023) Image fusion using dual tree discrete wavelet transform and weights optimization. Vis Comput 39(3):1181–1191
Babu BS, Narayana MV (2023) Two stage multi-modal medical image fusion with marine predator algorithm-based cascaded optimal dtcwt and nsst with deep learning. Biomed Signal Process Control 85:104921
Tan W, Tiwari P, Pandey HM, Moreira C, Jaiswal AK (2020) Multimodal medical image fusion algorithm in the era of big data. Neural Comput Appl 1–21
Li X, Zhou F, Tan H (2021) Joint image fusion and denoising via three-layer decomposition and sparse representation. Knowl Based Syst 224:107087
Li H, Xu T, Wu XJ, Lu J, Kittler J (2023) Lrrnet: A novel representation learning guided fusion network for infrared and visible images. IEEE Trans Pattern Anal Mach Intell
Meng L, Guo X, Li H (2019) Mri/ct fusion based on latent low rank representation and gradient transfer. Biomed Signal Process Control 53:101536
Li W, Peng X, Fu J, Wang G, Huang Y, Chao F (2022) A multiscale double-branch residual attention network for anatomical-functional medical image fusion. Comput Biol Med 141:105005
Xu H, Ma J, Jiang J, Guo X, Ling H (2020) U2fusion: A unified unsupervised image fusion network. IEEE Trans. Pattern Anal Mach Intell 44(1):502–518
Zhang H, Ma J (2021) Sdnet: A versatile squeeze-and-decomposition network for real-time image fusion. Int J Comput Vis 129(10):2761–2785
Zhang H, Xu H, Xiao Y, Guo X, Ma J (2020) Rethinking the image fusion: A fast unified image fusion network based on proportional maintenance of gradient and intensity. Proceedings of the AAAI Conference on Artificial Intelligence 34:12797–12804
Ma J, Xu H, Jiang J, Mei X, Zhang XP (2020) Ddcgan: A dual-discriminator conditional generative adversarial network for multi-resolution image fusion. IEEE Trans Image Process 29:4980– 4995
Fu J, Li W, Du J, Xu L (2021) Dsagan: A generative adversarial network based on dual-stream attention mechanism for anatomical and functional image fusion. Inf Sci 576(9)
Huang J, Le Z, Ma Y, Fan F, Zhang H, Yang L (2020) Mgmdcgan: Medical image fusion using multi-generator multi-discriminator conditional generative adversarial network. IEEE Access 8:55145–55157
Li X, Guo X, Han P, Wang X, Li H, Luo T (2020) Laplacian redecomposition for multimodal medical image fusion. IEEE Trans Instrum Meas 69(9):6880–6890
Liu X, Zhang B, Li X, Liu S, Yue C, Liang SY (2023) An approach for tool wear prediction using customized densenet and gru integrated model based on multi-sensor feature fusion. J Intell Manuf 34(2):885–902
Zhang G, Nie R, Cao J, Chen L, Zhu Y (2023) Fdgnet: A pair feature difference guided network for multimodal medical image fusion. Biomed Signal Process Control 81:104545
Zhang B, Wang Y, Ding C, Deng Z, Li L, Qin Z, Ding Z, Bian L, Yang C (2023) Multi-scale feature pyramid fusion network for medical image segmentation. Int J Comput Assist Radiol Surg 18(2):353–365
Liu L, Zhou Y, Huo J, Wu Y, Gu R (2023) Heterogenous image fusion model with sr-dual-channel pcnn significance region for nsst in an apple orchard. Appl Intell 1–22
Goyal S, Singh V, Rani A, Yadav N (2020) Fprsgf denoised non-subsampled shearlet transform-based image fusion using sparse representation. Signal Image Video Process 14:719–726
Dinh PH (2021) Multi-modal medical image fusion based on equilibrium optimizer algorithm and local energy functions. Appl Intell 51(11):8416–8431
Kolda TG, Bader BW (2009) Tensor decompositions and applications. SIAM review 51(3):455–500
Fu Z, Zhao Y, Chang D, Wang Y, Wen J (2022) Latent low-rank representation with weighted distance penalty for clustering. IEEE Trans Cybern 1–13. https://doi.org/10.1109/TCYB.2022.3166545
Zhao X, Yu Y, Zhou G, Zhao Q, Sun W (2022) Fast hypergraph regularized nonnegative tensor ring decomposition based on low-rank approximation. Appl Intell 1–24
Wang B, Niu H, Zeng J, Bai G, Lin S, Wang Y (2021) Latent representation learning model for multi-band images fusion via low-rank and sparse embedding. IEEE Trans Multimedia 23:3137–3152. https://doi.org/10.1109/TMM.2020.3020695
Woo S, Park J, Lee JY, Kweon IS (2018) Cbam: Convolutional block attention module. In: Proceedings of the European Conference on Computer Vision (ECCV), pp. 3–19
Hu J, Shen L, Sun G (2018) Squeeze-and-excitation networks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp.7132–7141
Song Q, Li J, Li C, Guo H, Huang R (2022) Fully attentional network for semantic segmentation. Proceedings of the AAAI Conference on Artificial Intelligence 36:2280–2288
Chen W, Zhu X, Sun R, He J, Li R, Shen X, Yu B (2020) Tensor low-rank reconstruction for semantic segmentation. In: European Conference on Computer Vision, pp. 52–69 Springer
Senhaji K, Ramchoun H, Ettaouil M (2020) Training feedforward neural network via multiobjective optimization model using non-smooth l1/2 regularization. Neurocomputing 410:1–11
Mo Y, Wu Y, Yang X, Liu F, Liao Y (2022) Review the state-of-the-art technologies of semantic segmentation based on deep learning. Neurocomputing 493:626–646
Cheng S, Wang Y, Huang H, Liu D, Liu S (2020) Nbnet: Noise basis learning for image denoising with subspace projection
Inanici MN, Navvab M (2006) The virtual lighting laboratory: Per-pixel luminance data analysis. Leukos 3(2):89–104
Wang Z, Bovik AC, Sheikh HR, Simoncelli EP (2004) Image quality assessment: from error visibility to structural similarity. IEEE transactions on image processing 13(4):600–612
Jung H, Kim Y, Jang H, Ha N, Sohn K (2020) Unsupervised deep image fusion with structure tensor representations. IEEE Trans Image Process 29:3845–3858
Bhandari M, Parajuli P, Chapagain P, Gaur L (2021) Evaluating performance of adam optimization by proposing energy index. In: International Conference on Recent Trends in Image Processing and Pattern Recognition, pp. 156–168 Springer
Piella G, Heijmans H (2003) A new quality metric for image fusion. In: Proceedings 2003 International Conference on Image Processing (Cat. No. 03CH37429), vol. 3, p. 173 IEEE
Han Y, Cai Y, Cao Y, Xu X (2013) A new image fusion performance metric based on visual information fidelity. Inf Fusion 14(2):127–135
Zhang L, Zhang L, Mou X, Zhang D (2011) Fsim: A feature similarity index for image quality assessment. IEEE Trans Image Process 20(8):2378–2386
Hossny M, Nahavandi S, Creighton D (2008) Comments on’information measure for performance of image fusion. Electron Lett 44(18):1066–1067
Qu G, Zhang D, Yan P (2002) Information measure for performance of image fusion. Electron Lett 38(7):1
Acknowledgements
This work was supported by the National Natural Science Foundation of China under Grants 61966037, 61833005, and 61463052, the China Postdoctoral Science Foundation under Grant 2017M621586, the Program of Yunnan Key Laboratory of Intelligent Systems and Computing (202205AG070003), and the Postgraduate Science Foundation of Yunnan University under Grants 2021Y263 and ZC-22222078. Key Project of Yunnan Basic Research Program under grant 202301AS070025 and Project Fund of Yunnan Provincial Department of Science and Technology 202105AF150011.
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Competing interest
The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
About this article
Cite this article
Pan, L., Nie, R., Zhang, G. et al. WAE-TLDN: self-supervised fusion for multimodal medical images via a weighted autoencoder and a tensor low-rank decomposition network. Appl Intell 54, 1656–1671 (2024). https://doi.org/10.1007/s10489-023-05097-z
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10489-023-05097-z