Skip to main content
Log in

Multi-focus image fusion via NSST with non-fixed base dictionary learning

  • Original Article
  • Published:
International Journal of System Assurance Engineering and Management Aims and scope Submit manuscript

Abstract

Multi-focus image fusion method can fuse images taken from the same view point with different focal settings, and obtain an image with every object in focus. In this paper, a novel multi-focus image fusion via non-subsampled shearlet transform (NSST) with non-fixed base dictionary learning is presented. First, low frequency coefficients and high frequency coefficients are obtained by NSST. Then, a new strategy, which can enhance the information of spatial detail for the fused image is proposed to process two different coefficients. The low frequency coefficients are fused via a non-fixed base dictionary, which makes the K-SVD algorithm more efficient, and the high frequency coefficients are fused with spatial frequency, which is effective in the fused image. Experiment results demonstrate that the results of proposed method obtain more spatial details and have almost zero residuals compared with several conventional methods in terms of both visual quality and objective measurements.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3

Similar content being viewed by others

References

  • Bengio Y, Louradour J, Collobert R, Weston J (2009) Curriculum learning. In: Proceedings of the 26th annual international conference on machine learning, pp 41–48

  • Bruckstein AM, Donoho DL, Elad M (2009) From sparse solutions of systems of equations to sparse modeling of signals and images. SIAM Rev 51:34–81

    Article  MathSciNet  Google Scholar 

  • Burt P, Adelson E (1983) The Laplacian pyramid as a compact image code. IEEE Trans Commun 31:532–540

    Article  Google Scholar 

  • Do MN, Vetterli M (2005) The contourlet transform: an efficient directional multiresolution image representation. IEEE Trans Image Process 14:2091–2106

    Article  Google Scholar 

  • Gao G, Xu L, Feng D (2013) Multi-focus image fusion based on non-subsampled shearlet transform. IET Image Process 7(6):633–639

    Article  Google Scholar 

  • Guo L, Dai M, Zhu M (2012) Multifocus color image fusion based on quaternion curvelet transform. Opt Express 20:18846

    Article  Google Scholar 

  • Li H, Qiu H, Yu Z, Li B (2017a) Multifocus image fusion via fixed window technique of multiscale images and non-local means filtering. Signal Process 138:71–85

    Article  Google Scholar 

  • Li S, Kang X, Fang L, Hu J, Yin H (2017b) Pixel-level image fusion: a survey of the state of the art. Inf Fusion 33:100–112

    Article  Google Scholar 

  • Liu Y, Wang Z (2015) Simultaneous image fusion and denosing with adaptive sparse representation. IET Image Process 9(5):347–357

    Article  Google Scholar 

  • Liu Y, Liu S, Wang Z (2015) A general framework for image fusion based on multi-scale transform and sparse representation. Inf Fusion 24:147–164

    Article  Google Scholar 

  • Nejati M, Samavi S, Shirani S (2015) Multi-focus image fusion using dictionary-based sparse representation. Inf Fusion 25:72–84

    Article  Google Scholar 

  • Qu G, Zhang D, Yan P (2002) Information measure for performance of image fusion. Electron Lett 38:313–315

    Article  Google Scholar 

  • Savic S (2011) Multifocus image fusion based on empirical mode decomposition. In: Twentieth international electro technical and computer science conference, 2011

  • Toet A (1989) Image fusion by a ratio of low-pass pyramid. Pattern Recognit Lett 9:245–253

    Article  Google Scholar 

  • Xydeas C, Petrovic V (2000) Objective image fusion performance measure. Electron Lett 36:308–309

    Article  Google Scholar 

  • Yang B, Li S (2010) Multifocus image fusion and restoration with sparse representation. IEEE Trans Instrum Meas 59:884–892

    Article  Google Scholar 

  • Yang C, Zhang J, Wang X, Liu X (2008) A novel similarity based quality metric for image fusion. Inf Fusion 9:156–160

    Article  Google Scholar 

  • Yang Y, Song T, Huang S, Lin P (2015) Multifocus image fusion based on NSCT and focused area detection. IEEE Sens J 15:2824–2838

    Google Scholar 

  • Yang Y, Yang M, Huang S, Ding M, Sun J (2018) Robust sparse representation combined with adaptive PCNN for multifocus image fusion. IEEE Access 6:20139–20151

    Google Scholar 

  • Yang Y, Zhang Y, Wu J, Li L, Huang S (2019) Multi-focus image fusion based on a non-fixed-base dictionary and multi-measure optimization. IEEE Access 7:46376–46388

    Article  Google Scholar 

  • Yu B, Jia B, Ding L, Cai Z, Wu Q, Law R (2016) Hybrid dual-tree complex wavelet transform and support vector machine for digital multi-focus image fusion. Neurocomputing 182:1–9

    Article  Google Scholar 

Download references

Acknowledgements

This work is supported by the National Natural Science Foundation of China (Nos. 61662026, 61862030, and 61462031), by the Natural Science Foundation of Jiangxi Province (Nos. 20182BCB22006, 20181BAB202010, 20192ACB20002, and 20192ACBL21008), and by the Project of the Education Department of Jiangxi Province (Nos. GJJ170318, GJJ170312, and KJLD14031).

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Yong Yang.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Yang, Y., Zhang, Y., Huang, S. et al. Multi-focus image fusion via NSST with non-fixed base dictionary learning. Int J Syst Assur Eng Manag 11, 849–855 (2020). https://doi.org/10.1007/s13198-019-00887-6

Download citation

  • Received:

  • Revised:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s13198-019-00887-6

Keywords

Navigation