Skip to main content

Advertisement

Log in

An efficient Dense-Resnet for multimodal image fusion using medical image

  • Published:
Multimedia Tools and Applications Aims and scope Submit manuscript

Abstract

Nowadays, the visual content of various medical images is increased through multimodal image fusion to gather more information from medical images. The complementary information available in various modalities is merged to increase the visual content of the image for quick medical diagnosis. However, the resultant fused multi-modality images suffered from different issues, like texture distortion and gradient, mainly for the affected region. Thus, a hybrid deep learning model, Dense-ResNet is designed for the fusing of multimodal medical images from different modalities in this research work. The images from three different modalities are collected initially, which are individually pre-processed by using a median filter. The pre-processed image of the spatial domain is transformed into a spectral domain by applying Dual-Tree Complex Wavelet Transform (DTCWT), and the transformed image is segmented using Edge-Attention Guidance Network (ET-Net). Finally, the multimodal fusion of medical images is performed on the segmented images using the designed Dense-ResNet model. Moreover, the superiority of the designed model is validated, which shows that the designed Dense-ResNet model outperforms as compared with other existing multimodal medical image fusion approaches. The Dense-ResNet model achieved 0.402 Mean Square Error (MSE), 0.634 Root Mean Square Error (RMSE), and 47.136 dB Peak Signal to Noise Ratio (PSNR).

This is a preview of subscription content, log in via an institution to check access.

Access this article

Subscribe and save

Springer+ Basic
$34.99 /Month
  • Get 10 units per month
  • Download Article/Chapter or eBook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6
Fig. 7

Similar content being viewed by others

Explore related subjects

Discover the latest articles and news from researchers in related subjects, suggested using machine learning.

Data Availability

The dataset used for the manuscript is the BRATS 2020 dataset taken fromhttps://www.kaggle.com/datasets/awsaf49/brats2020-trainingdata?select=BraTS20+Training+Metadata.csv”.

References

  1. Kaur M, Singh D (2021) Multi-modality medical image fusion technique using multi-objective differential evolution based deep neural networks. J Ambient Intell Humaniz Comput 12:2483–2493

    Article  Google Scholar 

  2. Wang L, Zhang J, Liu Y, Mi J, Zhang J (2021) Multimodal medical image fusion based on Gabor representation combination of multi-CNN and fuzzy neural network. IEEE Access 9:67634–67647

    Article  Google Scholar 

  3. Chang L, Feng X, Zhu X, Zhang R, He R, Xu C (2019) CT and MRI image fusion based on multiscale decomposition method and hybrid approach. IET Image Proc 13(1):83–88

    Article  Google Scholar 

  4. Lou XC, Feng X (2021) Multimodal medical image fusion based on multiple latent low-rank representation, Comput Math Meth Med, 1–16

  5. Nandhini Abirami R, Durai Raj Vincent PM, Srinivasan K, Manic KS, Chang CY (2022) Multimodal medical image fusion of positron emission tomography and magnetic resonance imaging using generative adversarial networks. Behav Neurol

  6. Li Y, Zhao J, Lv Z, Pan Z (2021) Multimodal medical supervised image fusion method by CNN. Front Neurosci 15:638976

    Article  Google Scholar 

  7. Huang B, Yang F, Yin M, Mo X, Zhong C (2020) A review of multimodal medical image fusion techniques. Comput Math Meth Med

  8. Meher B, Agrawal S, Panda R, Abraham A (2019) A survey on region based image fusion methods. Information Fusion 48:119–132

    Article  Google Scholar 

  9. Almasri MM, Alajlan AM (2022) Artificial Intelligence-Based Multimodal Medical Image Fusion Using Hybrid S2 Optimal CNN. Electronics 11(14):2124

    Article  Google Scholar 

  10. Fu J, Li W, Du J, Xiao B (2020) Multimodal medical image fusion via laplacian pyramid and convolutional neural network reconstruction with local gradient energy strategy. Comput Biol Med 126:104048

    Article  Google Scholar 

  11. Hill P, Al-Mualla ME, Bull D (2016) Perceptual image fusion using wavelets. IEEE Trans Image Process 26(3):1076–1088

    Article  MathSciNet  Google Scholar 

  12. Li S, Yin H, Fang L (2012) Group-sparse representation with dictionary learning for medical image denoising and fusion. IEEE Trans Biomed Eng 59(12):3450–3459

    Article  Google Scholar 

  13. Liu Y, Liu S, Wang Z (2015) A general framework for image fusion based on multi-scale transform and sparse representation. Information fusion 24:147–164

    Article  Google Scholar 

  14. Li Y, Zhao J, Lv Z, Li J (2021) Medical image fusion method by deep learning. Intl J Cognit Comput Eng 2:21–29

    Article  Google Scholar 

  15. Liu Y, Chen X, Peng H, Wang Z (2017) Multi-focus image fusion with a deep convolutional neural network. Information Fusion 36:191–207

    Article  Google Scholar 

  16. Li W, Piëch V, Gilbert CD (2004) Perceptual learning and top-down influences in primary visual cortex. Nat Neurosci 7(6):651–657

    Article  Google Scholar 

  17. Gilbert CD, Li W (2013) Top-down influences on visual processing. Nat Rev Neurosci 14(5):350–363

    Article  Google Scholar 

  18. Tang W, He F, Liu Y, Duan Y (2022) MATR: multimodal medical image fusion via multiscale adaptive transformer. IEEE Trans Image Process 31:5134–5149

    Article  Google Scholar 

  19. Maheshan CM, Prasanna Kumar H (2020) Performance of image pre-processing filters for noise removal in transformer oil images at different temperatures. SN Appl Sci, 2: 1–7

  20. Slimen IB, Boubchir L, Mbarki Z, Seddik H (2020) EEG epileptic seizure detection and classification based on dual-tree complex wavelet transform and machine learning algorithms. J Biomed Res 34(3):151

    Article  Google Scholar 

  21. Zhang Z, Fu H, Dai H, Shen J, Pang Y, Shao L (2019) Et-net: A generic edge-attention guidance network for medical image segmentation. In Proceedings of 22nd International Conference, In Medical Image Computing and Computer Assisted Intervention–MICCAI 2019, Shenzhen, China, Part I vol.22, Springer International Publishing, pp. 442–450, October 13–17

  22. Vulli A, Srinivasu PN, Sashank MSK, Shafi J, Choi J, Ijaz MF (2022) Fine-tuned DenseNet-169 for breast cancer metastasis prediction using FastAI and 1-cycle policy. Sensors 22(8):2988

    Article  Google Scholar 

  23. Chen Z, Chen Y, Wu L, Cheng S, Lin P (2019) Deep residual network based fault detection and diagnosis of photovoltaic arrays using current-voltage curves and ambient conditions. Energy Convers Manage 198:111793

    Article  Google Scholar 

  24. Bhaladhare PR, Jinwala DC (2014) A clustering approach for the-diversity model in privacy preserving data mining using fractional calculus-bacterial foraging optimization algorithm. Adv Comput Eng

  25. BRATS 2020 dataset is taken from “https://www.kaggle.com/datasets/awsaf49/brats2020-training-data?select=BraTS20+Training+Metadata.csv”, accessed on July 2023

  26. Tang W, Fazhi F, Liu Y, Duan Y (2022) MATR: Multimodal Medical Image Fusion via Multiscale Adaptive Transformer. IEEE Trans Image Process 31:5134–5149

    Article  Google Scholar 

  27. Weiwei Kong, Chi Li, Yang Lei (2022) Multimodal medical image fusion using convolutional neural network and extreme learning machine. Front Neurorob, Vol. 16, November

Download references

Acknowledgements

I would like to express my very great appreciation to the co-authors of this manuscript for their valuable and constructive suggestions during the planning and development of this research work.

Funding

This research did not receive any specific funding.

Author information

Authors and Affiliations

Authors

Contributions

All authors have made substantial contributions to conception and design, revising the manuscript, and the final approval of the version to be published. Also, all authors agreed to be accountable for all aspects of the work in ensuring that questions related to the accuracy or integrity of any part of the work are appropriately investigated and resolved.

Corresponding author

Correspondence to Tanima Ghosh.

Ethics declarations

Conflict of interest

The authors declare no conflict of interest.

Informed consent

Not Applicable.

Ethical approval

Not Applicable.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Ghosh, T., Jayanthi, N. An efficient Dense-Resnet for multimodal image fusion using medical image. Multimed Tools Appl 83, 68181–68208 (2024). https://doi.org/10.1007/s11042-024-18974-7

Download citation

  • Received:

  • Revised:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s11042-024-18974-7

Keywords