Skip to main content
Log in

MBRARN: multibranch residual attention reconstruction network for medical image fusion

  • Original Article
  • Published:
Medical & Biological Engineering & Computing Aims and scope Submit manuscript

Abstract

Medical image fusion aims to integrate complementary information from multimodal medical images and has been widely applied in the field of medicine, such as clinical diagnosis, pathology analysis, and healing examinations. For the fusion task, feature extraction is a crucial step. To obtain significant information embedded in medical images, many deep learning-based algorithms have been proposed recently and achieved good fusion results. However, most of them can hardly capture the independent and underlying features, which leads to unsatisfactory fusion results. To address these issues, a multibranch residual attention reconstruction network (MBRARN) is proposed for the medical image fusion task. The proposed network mainly consists of three parts: feature extraction, feature fusion, and feature reconstruction. Firstly, the input medical images are converted into three scales by image pyramid operation and then are input into three branches of the proposed network respectively. The purpose of this procedure is to capture the local detailed information and the global structural information. Then, convolutions with residual attention modules are designed, which can not only enhance the captured outstanding features, but also make the network converge fast and stably. Finally, feature fusion is performed with the designed fusion strategy. In this step, a new more effective fusion strategy is correspondently designed for MRI-SPECT based on the Euclidean norm, called feature distance ratio (FDR). The experimental results conducted on Harvard whole brain atlas dataset demonstrate that the proposed network can achieve better results in terms of both subjective and objective evaluation, compared with some state-of-the-art medical image fusion algorithms.

Graphical Abstract

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Algorithm 1
Fig. 5
Fig. 6
Fig. 7
Fig. 8
Fig. 9
Fig. 10
Fig. 11
Fig. 12
Fig. 13
Fig. 14
Fig. 15

Similar content being viewed by others

Data availability

All data included in this study are available upon request by contact with the corresponding author.

References

  1. Kang TW, Lee MW, Song KD et al (2017) Added value of contrast-enhanced ultrasound on biopsies of focal hepatic lesions invisible on fusion imaging guidance. Korean J Radiol 18(1):152–161

    Article  PubMed  PubMed Central  Google Scholar 

  2. Wu XB, Li KY, Luo HC et al (2018) The diagnostic value of focal liver lesion (≤2 cm) undetectable on conventional ultrasound by image fusion with contrast-enhanced ultrasound. Chin J Ultrason 27(10):860–864

    Google Scholar 

  3. Hermessi H, Mourali O, Zagrouba E (2021) Multimodal medical image fusion review: theoretical background and recent advances. Signal Process 183:108036

    Article  Google Scholar 

  4. Tawfik N, Elnemr HA, Fakhr M et al (2021) Survey study of multimodality medical image fusion methods. Multimedia Tools Appl 80:6369–6396

    Article  Google Scholar 

  5. Zhou T, Cheng QR, Lu HL et al (2023) Deep learning methods for medical image fusion: a review. Comput Biol Med 160:106959

    Article  PubMed  Google Scholar 

  6. Dong Y, Wang WP, Mao F et al (2016) Application of imaging fusion combining contrast-enhanced ultrasound and magnetic resonance imaging in detection of hepatic cellular carcinomas undetectable by conventional ultrasound. J Gastroenterol Hepatol 31(4):822–828

    Article  PubMed  Google Scholar 

  7. Wu DF, He W, Lin S et al (2018) The application of volume navigation with ultrasound and MR fusion image in neurosurgical brain tumor resection. Chinese J Ultrason 27(12):1036–1041

    Google Scholar 

  8. Zhang SY, Lv B (2023) A review of the diagnostic value of ultrasound image fusion in liver space occupying lesions. J Shandong First Med Un Shandong Acad Med Sci 44(4):313–316

    Google Scholar 

  9. Ma HL, Kui GH, Zhang Z (2020) Value of multi-modality medical image fusion technology in evaluating hepatic focal lesions in chronic aplastic anemia patients administered androgens. J Clin Exp Med 19(22):2456–2459

    Google Scholar 

  10. Ma BR, Yang H (2005) Image fusion technology and its application in epileptic. Chin J Biomed Eng 24(3):357–361

    Google Scholar 

  11. Hu J, Dong XQ, Lin Q et al (2018) Medical image registration and fusion technology applied in cancer radiotherapy. Chin Med Equip J 39(8):75–7884

    CAS  Google Scholar 

  12. Du J, LI W, XIAO B et al (2016) Union Laplacian pyramid with multiple features for medical image fusion. Neurocomputing 194:326–339

    Article  Google Scholar 

  13. Wang Z, Cui Z, Zhu Y (2020) Multi-modal medical image fusion by Laplacian pyramid and adaptive sparse representation. Comput Biol Med 123:103823

    Article  PubMed  Google Scholar 

  14. Bhat S, Koundal D (2021) Multi-focus image fusion using neutrosophic based wavelet transform. Appl Soft Comput 106:107307

    Article  Google Scholar 

  15. Liu Y, Liu S, Wang Z (2014) Medical image fusion by combining nonsubsampled contourlet transform and sparse representation. In: Pattern Recognition: 6th Chinese Conference (CCPR). pp 17–19

  16. Ibrahim SI, Makhlouf MA, El-Tawel GS (2023) Multimodal medical image fusion algorithm based on pulse coupled neural networks and nonsubsampled contourlet transform. Med Biol Eng Compu 61(1):155–177

    Article  Google Scholar 

  17. Xu W, Fu YL, Xu H (2023) Medical image fusion using enhanced cross-visual cortex model based on artificial selection and impulse-coupled neural network. Comput Methods Programs Biomed 229:107304

    Article  PubMed  Google Scholar 

  18. Yin M, Liu X, Liu Y et al (2018) Medical image fusion with parameter-adaptive pulse coupled neural network in nonsubsampled shearlet transform domain. IEEE Trans Instrum Meas 68(1):49–64

    Article  Google Scholar 

  19. Wang L, Dou J, Qin P et al (2021) Multimodal medical image fusion based on nonsubsampled shearlet transform and convolutional sparse representation. Multimedia Tools Appl 80:36401–36421

    Article  Google Scholar 

  20. Gai D, Shen X, Cheng H et al (2019) Medical image fusion via PCNN based on edge preservation and improved sparse representation in NSST domain. IEEE Access 7:85413–85429

    Article  Google Scholar 

  21. Dinh PH (2023) Combining spectral total variation with dynamic threshold neural P systems for medical image fusion. Biomed Signal Process Control 80:104343

    Article  Google Scholar 

  22. Li J, Han D, Wang X et al (2023) Multi-sensor medical-image fusion technique based on embedding bilateral filter in least squares and salient detection. Sensors 23(7):3490

    Article  PubMed  PubMed Central  Google Scholar 

  23. Dinh PH (2023) Medical image fusion based on enhanced three-layer image decomposition and Chameleon swarm algorithm. Biomed Signal Process Control 84:104740

    Article  Google Scholar 

  24. Barba-J L, Vargas-Quintero L, Calderón-Agudelo JA (2022) Bone SPECT/CT image fusion based on the discrete hermite transform and sparse representation. Biomed Signal Process Control 71:103096

    Article  Google Scholar 

  25. Chen J, Zhang L, Lu L et al (2021) A novel medical image fusion method based on rolling guidance filtering. Int Things 14:100172

    Article  Google Scholar 

  26. Zhao H, Zhang JX, Zhang ZG (2021) PCNN medical image fusion based on NSCT and DWT. Adv Laser Optoelectron 58(20):445–454

    Google Scholar 

  27. Liu Y, Chen X, Cheng J et al (2017) A medical image fusion method based on convolutional neural networks[C]//2017 20th international conference on information fusion (Fusion). IEEE 2017:1–7

    Google Scholar 

  28. Wang L, Zhang J, Liu Y et al (2021) Multimodal medical image fusion based on Gabor representation combination of multi-CNN and fuzzy neural network. IEEE Access 9:67634–67647

    Article  Google Scholar 

  29. Zhang Y, Liu Y, Sun P et al (2020) IFCNN: A general image fusion framework based on convolutional neural network. Inform Fusion 54:99–118

    Article  Google Scholar 

  30. Xu H, Ma J (2021) EMFusion: An unsupervised enhanced medical image fusion network. Inform Fusion 76:177–186

    Article  Google Scholar 

  31. Azam MA, Khan KB, Salahuddin S et al (2022) A review on multimodal medical image fusion: compendious analysis of medical modalities, multimodal databases, fusion techniques and quality metrics. Comput Biol Med 144:105253

    Article  PubMed  Google Scholar 

  32. Zhang G, Nie R, Cao J et al (2023) FDGNet: A pair feature difference guided network for multimodal medical image fusion. Biomed Signal Process Control 81:104545

    Article  Google Scholar 

  33. Fu J, He B, Yang J et al (2023) CDRNet: Cascaded dense residual network for grayscale and pseudocolor medical image fusion. Comput Methods Programs Biomed 234:107506

    Article  PubMed  Google Scholar 

  34. Fu J, Li W, Peng X et al (2023) MDRANet: A multiscale dense residual attention network for magnetic resonance and nuclear medicine image fusion. Biomed Signal Process Control 80:104382

    Article  Google Scholar 

  35. Ding ZS, Li HY, Guo Y et al (2023) M4FNet: Multimodal medical image fusion network via multi-receptive-field and multi-scale feature integration. Comput Biol Med 159:106923

    Article  PubMed  Google Scholar 

  36. Panigrahy C, Seal A, Gonzalo-Martín C et al (2023) Parameter adaptive unit-linking pulse coupled neural network based MRI-PET/SPECT image fusion. Biomed Signal Process Control 83:104659

    Article  Google Scholar 

  37. Zhao C, Wang T, Lei B (2021) Medical image fusion method based on dense block and deep convolutional generative adversarial network. Neural Comput Appl 33:6595–6610

    Article  Google Scholar 

  38. Huang J, Le Z, Ma Y et al (2020) MGMDcGAN: medical image fusion using multi-generator multi-discriminator conditional generative adversarial network. IEEE Access 8:55145–55157

    Article  Google Scholar 

  39. Ma J, Xu H, Jiang J et al (2020) DDcGAN: A dual-discriminator conditional generative adversarial network for multi-resolution image fusion. IEEE Trans Image Process 29:4980–4995

    Article  Google Scholar 

  40. Fu J, Li W, Du J et al (2021) DSAGAN: A generative adversarial network based on dual-stream attention mechanism for anatomical and functional image fusion. Inf Sci 576:484–506

    Article  Google Scholar 

  41. Wang J, Yu L, Tian S (2022) MsRAN: a multi-scale residual attention network for multi-model image fusion. Med Biol Eng Compu 60:3615–3634

    Article  Google Scholar 

  42. He K, Zhang X, Ren S, et al (2016) Deep residual learning for image recognition. In: Proceedings of the IEEE conference on computer vision and pattern recognition. pp 770–778

  43. Mnih V, Heess N, Graves A (2014) Recurrent models of visual attention. In: Advances in neural information processing systems

  44. Wang F, Jiang M, Qian C, et al (2017) Residual attention network for image classification. In: Proceedings of the IEEE conference on computer vision and pattern recognition. pp 3156–3164

  45. Kingma DP, Ba J (2014) Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980. https://doi.org/10.48550/arXiv.1412.6980

  46. Tang LF, Zhang H, Xu H, Ma JY (2023) Deep learning-based image fusion: a survey. J Image Graphics 28(1):3–36

    Google Scholar 

  47. Wang Z, Bovik AC, Sheikh HR, Simoncelli EP (2004) Image quality assessment: from error visibility to structural similarity. IEEE Trans Image Process 13(4):600–612

    Article  PubMed  Google Scholar 

  48. Wang Z, Li Q (2011) Information content weighting for perceptual image quality assessment. IEEE Trans Image Process 20(5):1185–1198

    Article  PubMed  Google Scholar 

  49. Qu G, Zhang D, Yan P (2002) Information measure for performance of image fusion. Electron Lett 38(7):313–315

    Article  Google Scholar 

  50. Haghighat MBA, Aghagolzadeh A, Seyedarabi H (2011) A non-reference image fusion metric based on mutual information of image features. Comput Electr Eng 37(5):744–756

    Article  Google Scholar 

  51. Sheikh HR, Bovik AC (2016) Image information and visual quality. IEEE Trans Image Process 15(2):430–444

    Article  Google Scholar 

  52. Xydeas CS, Petrovic V (2000) Objective image fusion performance measure. Electron Lett 36(4):308–309

    Article  Google Scholar 

  53. Hossny M, Nahavandi S, Creighton D et al (2010) Image fusion performance metric based on mutual information and entropy driven quadtree decomposition. Electron Lett 46(18):1266–1268

    Article  Google Scholar 

  54. Aslantas V, Bendes E (2015) A new image quality metric for image fusion: the sum of the correlations of differences. Aeu-int J Elect Commun 69(12):1890–1896

    Article  Google Scholar 

  55. Zhu Z, Zheng M, Qi G et al (2019) A phase congruency and local Laplacian energy based multi-modality medical image fusion method in NSCT domain. IEEE Access 7:20811–20824

    Article  Google Scholar 

  56. Tan W, Zhang J, Xiang P et al (2020) Infrared and visible image fusion via NSST and PCNN in multiscale morphological gradient domain[C]//Optics, photonics and digital technologies for imaging applications VI. SPIE 11353:297–303

    Google Scholar 

  57. Tan W, Thitøn W, Xiang P et al (2021) Multi-modal brain image fusion based on multi-level edge-preserving filtering. Biomed Signal Process Control 64:102280

    Article  Google Scholar 

  58. Liu Y, Liu S, Wang Z (2015) A general framework for image fusion based on multi-scale transform and sparse representation. Inform Fusion 24:147–164

    Article  Google Scholar 

  59. Vanitha K, Satyanarayana D, Prasad MNG (2020) Medical image fusion algorithm based on weighted local energy motivated PAPCNN in NSST domain. J Adv Res Dyn Control Syst 12(SP3):960–967

    Article  Google Scholar 

  60. Khorasani A, Tavakoli MB, Saboori M et al (2021) Preliminary study of multiple b-value diffusion-weighted images and T1 post enhancement magnetic resonance imaging images fusion with Laplacian Re-decomposition (LRD) medical fusion algorithm for glioma grading. Eur J Radiol Open 8:100378

    Article  PubMed  PubMed Central  Google Scholar 

  61. Li W, Peng X, Fu J et al (2022) A multiscale double-branch residual attention network for anatomical–functional medical image fusion. Comput Biol Med 141:105005

    Article  PubMed  Google Scholar 

Download references

Acknowledgements

The authors would like to thank the anonymous reviewers for their critical and constructive comments and suggestions.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Lei Yu.

Ethics declarations

Conflict of interest

The authors declare no competing interests.

Additional information

Publisher's note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Zhang, W., Lu, Y., Zheng, H. et al. MBRARN: multibranch residual attention reconstruction network for medical image fusion. Med Biol Eng Comput 61, 3067–3085 (2023). https://doi.org/10.1007/s11517-023-02902-2

Download citation

  • Received:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s11517-023-02902-2

Keywords

Navigation