Skip to main content
Log in

ICWGAN-GP: an image fusion method based on infrared compensator and wasserstein generative adversarial network with gradient penalty

  • Published:
Applied Intelligence Aims and scope Submit manuscript

Abstract

The existing Generative adversarial network (GAN)-based infrared (IR) and visible (VIS) image fusion methods mainly used multiple discriminators to preserve salient information in source images, which brings difficulty in balancing the performance of these discriminators during training, leading to unideal fused results. To tackle this disadvantage, an image fusion method based on IR compensator and Wasserstein generative adversarial network with gradient penalty (WGAN-GP) is proposed, called ICWGAN-GP. The generator of ICWGAN-GP employs an adjustment mechanism to obtain more VIS gradients while getting IR intensities, and important details in VIS images are highlighted through the adversarial game between a discriminator and a generator. Using one discriminator allows ICWGAN-GP to focus on learning the feature distribution in a source image, which avoids the balance problem caused by multiple discriminators, and improves the efficiency of the ICWGAN-GP. In addition, an IR compensator based on Quadtree-Bézier method is designed to make up for bright IR features in the fused images. Extensive experiments on public datasets show that ICWGAN-GP can highlight bright target features while generating rich texture in the fused images, and achieves better objective metrics in terms of SCD, CC, FMI_W and VIF than the state-of-the-art methods like U2Fusion, MDLatLRR, DDcGAN, etc. Moreover, in our further fusion tracking experiments, ICWGAN-GP also demonstrates good tracking performance.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6
Algorithm 1
Fig. 7
Fig. 8
Fig. 9
Fig. 10
Fig. 11
Fig. 12
Fig. 13

Similar content being viewed by others

Data availability statement

The datasets generated during and/or analysed during the current study are available from the corresponding author on reasonable request.

References

  1. Ma J, Ma Y, Li C (2019) Infrared and visible image fusion methods and applications: A survey. Inf. Fusion 45, 153–178 (2019-01-01). https://doi.org/10.1016/j.inffus.2018.02.004. https://www.sciencedirect.com/science/article/pii/S1566253517307972

  2. Ren L, Pan Z, Cao J, Zhang H, Wang H (2021) Infrared and visible image fusion based on edge-preserving guided filter and infrared feature decomposition. Signal Process. 186(108):108

    Google Scholar 

  3. Chen J, Li X, Luo L, Mei X, Ma J (2020) Infrared and visible image fusion based on target-enhanced multiscale transform decomposition. Inf Sci 508:64–78

    Article  Google Scholar 

  4. Li H, He X, Yu Z, Luo J (2020) Noise-robust image fusion with low-rank sparse decomposition guided by external patch prior. Inf Sci 523:14–37

    Article  MathSciNet  MATH  Google Scholar 

  5. Chen J, Wu K, Cheng Z, Luo L (2021) A saliency-based multiscale approach for infrared and visible image fusion. Signal Process. (182): 107,936 https://doi.org/10.1016/j.sigpro.2020.107936. https://www.sciencedirect.com/science/article/pii/S0165168420304801

  6. Li H, Wu XJ (2019) DenseFuse: A Fusion Approach to Infrared and Visible Images. IEEE Trans. Image Process. 28(5):2614–2623. https://doi.org/10.1109/TIP.2018.2887342

    Article  MathSciNet  Google Scholar 

  7. Li H, Wu XJ, Durrani T (2020) NestFuse: An Infrared and Visible Image Fusion Architecture Based on Nest Connection and Spatial/Channel Attention Models. IEEE Trans Instrum Meas 69(12):9645–9656. https://doi.org/10.1109/TIM.2020.3005230

    Article  Google Scholar 

  8. Li H, Wu XJ, Kittler J (2021) RFN-Nest: An end-to-end residual fusion network for infrared and visible images. Inf Fusion 73:72–86. https://doi.org/10.1016/j.inffus.2021.02.023. https://www.sciencedirect.com/science/article/pii/S1566253521000440

  9. Vs V, Valanarasu JMJ, Oza P, Patel VM (2022) Image fusion transformer. In: 2022 IEEE International Conference on Image Processing (ICIP), pp. 3566–3570. IEEE

  10. Ma J, Tang L, Fan F, Huang J, Mei X, Ma Y (2022) Swinfusion: Cross-domain long-range learning for general image fusion via swin transformer. IEEE/CAA J Automatica Sinica 9(7):1200–1217

    Article  Google Scholar 

  11. Ma J, Yu W, Liang P, Li C, Jiang J (2019) FusionGAN: A generative adversarial network for infrared and visible image fusion. Inf Fusion 48:11–26. https://doi.org/10.1016/j.inffus.2018.09.004. https://www.sciencedirect.com/science/article/pii/S1566253518301143

  12. Ma J, Xu H, Jiang J, Mei X, Zhang XP (2020) DDcGAN: A Dual-Discriminator Conditional Generative Adversarial Network for Multi-Resolution Image Fusion. IEEE Trans. Image Process. 29:4980–4995. https://doi.org/10.1109/TIP.2020.2977573

    Article  MATH  Google Scholar 

  13. Ma J, Liang P, Yu W, Chen C, Guo X, Wu J, Jiang J (2020) Infrared and visible image fusion via detail preserving adversarial learning. Inf Fusion 54:85–98. https://doi.org/10.1016/j.inffus.2019.07.005. https://www.sciencedirect.com/science/article/pii/S1566253519300314

  14. Zhang Y, Zhang L, Bai X, Zhang L (2017) Infrared and visual image fusion through infrared feature extraction and visual information preservation. Infrared Phys. Technol 83:227–237. https://doi.org/10.1016/j.infrared.2017.05.007. https://www.sciencedirect.com/science/article/pii/S1350449517300725

  15. Arjovsky M, Chintala S, Bottou L (2017) Wasserstein Generative Adversarial Networks. In: Proceedings of the 34th International Conference on Machine Learning, PMLR pp. 214–223. https://proceedings.mlr.press/v70/arjovsky17a.html

  16. Gulrajani I, Ahmed F, Arjovsky M, Dumoulin V, Courville AC (2017) Improved Training of Wasserstein GANs. In: Advances in Neural Information Processing Systems, vol. 30. Curran Associates, Inc.. https://proceedings.neurips.cc/paper/2017/hash/892c3b1c6dccd52936e27cbd0ff683d6-Abstract.html

  17. Zhang L (2008) In situ image segmentation using the convexity of illumination distribution of the light sources. IEEE transactions on pattern analysis and machine intelligence 30(10):1786–1799

    Article  Google Scholar 

  18. Shreyamsha Kumar BK (2015) Image fusion based on pixel significance using cross bilateral filter. Signal, Image and Video Processing 9(5):1193–1204. https://doi.org/10.1007/s11760-013-0556-9

    Article  Google Scholar 

  19. Ma J, Chen C, Li C, Huang J (2016) Infrared and visible image fusion via gradient transfer and total variation minimization. Inf Fusion 31:100–109. https://doi.org/10.1016/j.inffus.2016.02.001. https://www.sciencedirect.com/science/article/pii/S156625351630001X

  20. Bavirisetti DP, Xiao G, Liu G (2017) Multi-sensor image fusion based on fourth order partial differential equations. In: 2017 20th International Conference on Information Fusion (Fusion), pp. 1–9. https://doi.org/10.23919/ICIF.2017.8009719

  21. Xu H, Ma J, Jiang J, Guo X, Ling H (2022) U2Fusion: A Unified Unsupervised Image Fusion Network. IEEE Trans. Pattern Anal Mach Intell 44(1):502–518. https://doi.org/10.1109/TPAMI.2020.3012548

    Article  Google Scholar 

  22. Li H, Wu XJ, Kittler J (2020) MDLatLRR: A Novel Decomposition Method for Infrared and Visible Image Fusion. IEEE Trans. Image Process 29:4733–4746. https://doi.org/10.1109/TIP.2020.2975984

    Article  MATH  Google Scholar 

  23. Ma J, Zhang H, Shao Z, Liang P, Xu H (2021) GANMcC: A Generative Adversarial Network With Multiclassification Constraints for Infrared and Visible Image Fusion. IEEE Trans Instrum Meas 70:1–14. https://doi.org/10.1109/TIM.2020.3038013

  24. Meher B, Agrawal S, Panda R, Abraham A (2019) A survey on region based image fusion methods. Inf Fusion 48, 119–132 (2019-08-01). https://doi.org/10.1016/j.inffus.2018.07.010. https://www.sciencedirect.com/science/article/pii/S1566253517307583

  25. Han Y, Cai Y, Cao Y, Xu X (2013) A new image fusion performance metric based on visual information fidelity. Inf Fusion 14(2):127–135. https://doi.org/10.1016/j.inffus.2011.08.002. https://www.sciencedirect.com/science/article/pii/S156625351100056X

  26. Deshmukh M, Bhosale U, Deshmukh M, Engg SCO, Mumbai N, Bhosale U (2010) Image Fusion and Image Quality Assessment of Fused Images. International Journal of Image Processing (IJIP

  27. Wang Z, Bovik A, Sheikh H (2004) Simoncelli E (2004) Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process. 13(4):600–612. https://doi.org/10.1109/TIP.2003.819861

    Article  Google Scholar 

  28. Aslantas V (2015) Bendes E (2015) A new image quality metric for image fusion: The sum of the correlations of differences. AEU - International J Electron Commun 69(12):1890–1896. https://doi.org/10.1016/j.aeue.2015.09.004. https://www.sciencedirect.com/science/article/pii/S1434841115002691

  29. Haghighat M, Razian MA (2014) Fast-fmi: non-reference image fusion metric.In: 2014 IEEE 8th International Conference on Application of Information and Communication Technologies (AICT), IEEE pp. 1–3

  30. Zhang X, Ye P, Leung H, Gong K, Xiao G (2020) Object fusion tracking based on visible and infrared images: A comprehensive review. Inf Fusion 63:166–187. https://doi.org/10.1016/j.inffus.2020.05.002. https://www.sciencedirect.com/science/article/pii/S1566253520302657

  31. Yang T, Xu P, Hu R, Chai H, Chan AB (2020) ROAM: Recurrently Optimizing Tracking Model. pp. 6718–6727. https://openaccess.thecvf.com/content_CVPR_2020/html/Yang_ROAM_Recurrently_Optimizing_Tracking_Model_CVPR_2020_paper.html

Download references

Acknowledgements

The authors gratefully acknowledge the financial supports by The National Science Foundation of China(62203224), Shanghai Special Plan for Local Colleges and Universities for Capacity Building (22010501300).

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Gang Liu.

Ethics declarations

Conflicts of interest

We declare that we have no financial and personal relationships with other people or organizations that can inappropriately influence our work, there is no professional or other personal interest of any nature or kind in any product, service or company that could be construed as influencing the position presented in the manuscript entitled "Manuscript".

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Wang, X., Liu, G., Tang, L. et al. ICWGAN-GP: an image fusion method based on infrared compensator and wasserstein generative adversarial network with gradient penalty. Appl Intell 53, 27637–27654 (2023). https://doi.org/10.1007/s10489-023-04933-6

Download citation

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s10489-023-04933-6

Keywords

Navigation