Abstract
The existing Generative adversarial network (GAN)-based infrared (IR) and visible (VIS) image fusion methods mainly used multiple discriminators to preserve salient information in source images, which brings difficulty in balancing the performance of these discriminators during training, leading to unideal fused results. To tackle this disadvantage, an image fusion method based on IR compensator and Wasserstein generative adversarial network with gradient penalty (WGAN-GP) is proposed, called ICWGAN-GP. The generator of ICWGAN-GP employs an adjustment mechanism to obtain more VIS gradients while getting IR intensities, and important details in VIS images are highlighted through the adversarial game between a discriminator and a generator. Using one discriminator allows ICWGAN-GP to focus on learning the feature distribution in a source image, which avoids the balance problem caused by multiple discriminators, and improves the efficiency of the ICWGAN-GP. In addition, an IR compensator based on Quadtree-Bézier method is designed to make up for bright IR features in the fused images. Extensive experiments on public datasets show that ICWGAN-GP can highlight bright target features while generating rich texture in the fused images, and achieves better objective metrics in terms of SCD, CC, FMI_W and VIF than the state-of-the-art methods like U2Fusion, MDLatLRR, DDcGAN, etc. Moreover, in our further fusion tracking experiments, ICWGAN-GP also demonstrates good tracking performance.
Similar content being viewed by others
Data availability statement
The datasets generated during and/or analysed during the current study are available from the corresponding author on reasonable request.
References
Ma J, Ma Y, Li C (2019) Infrared and visible image fusion methods and applications: A survey. Inf. Fusion 45, 153–178 (2019-01-01). https://doi.org/10.1016/j.inffus.2018.02.004. https://www.sciencedirect.com/science/article/pii/S1566253517307972
Ren L, Pan Z, Cao J, Zhang H, Wang H (2021) Infrared and visible image fusion based on edge-preserving guided filter and infrared feature decomposition. Signal Process. 186(108):108
Chen J, Li X, Luo L, Mei X, Ma J (2020) Infrared and visible image fusion based on target-enhanced multiscale transform decomposition. Inf Sci 508:64–78
Li H, He X, Yu Z, Luo J (2020) Noise-robust image fusion with low-rank sparse decomposition guided by external patch prior. Inf Sci 523:14–37
Chen J, Wu K, Cheng Z, Luo L (2021) A saliency-based multiscale approach for infrared and visible image fusion. Signal Process. (182): 107,936 https://doi.org/10.1016/j.sigpro.2020.107936. https://www.sciencedirect.com/science/article/pii/S0165168420304801
Li H, Wu XJ (2019) DenseFuse: A Fusion Approach to Infrared and Visible Images. IEEE Trans. Image Process. 28(5):2614–2623. https://doi.org/10.1109/TIP.2018.2887342
Li H, Wu XJ, Durrani T (2020) NestFuse: An Infrared and Visible Image Fusion Architecture Based on Nest Connection and Spatial/Channel Attention Models. IEEE Trans Instrum Meas 69(12):9645–9656. https://doi.org/10.1109/TIM.2020.3005230
Li H, Wu XJ, Kittler J (2021) RFN-Nest: An end-to-end residual fusion network for infrared and visible images. Inf Fusion 73:72–86. https://doi.org/10.1016/j.inffus.2021.02.023. https://www.sciencedirect.com/science/article/pii/S1566253521000440
Vs V, Valanarasu JMJ, Oza P, Patel VM (2022) Image fusion transformer. In: 2022 IEEE International Conference on Image Processing (ICIP), pp. 3566–3570. IEEE
Ma J, Tang L, Fan F, Huang J, Mei X, Ma Y (2022) Swinfusion: Cross-domain long-range learning for general image fusion via swin transformer. IEEE/CAA J Automatica Sinica 9(7):1200–1217
Ma J, Yu W, Liang P, Li C, Jiang J (2019) FusionGAN: A generative adversarial network for infrared and visible image fusion. Inf Fusion 48:11–26. https://doi.org/10.1016/j.inffus.2018.09.004. https://www.sciencedirect.com/science/article/pii/S1566253518301143
Ma J, Xu H, Jiang J, Mei X, Zhang XP (2020) DDcGAN: A Dual-Discriminator Conditional Generative Adversarial Network for Multi-Resolution Image Fusion. IEEE Trans. Image Process. 29:4980–4995. https://doi.org/10.1109/TIP.2020.2977573
Ma J, Liang P, Yu W, Chen C, Guo X, Wu J, Jiang J (2020) Infrared and visible image fusion via detail preserving adversarial learning. Inf Fusion 54:85–98. https://doi.org/10.1016/j.inffus.2019.07.005. https://www.sciencedirect.com/science/article/pii/S1566253519300314
Zhang Y, Zhang L, Bai X, Zhang L (2017) Infrared and visual image fusion through infrared feature extraction and visual information preservation. Infrared Phys. Technol 83:227–237. https://doi.org/10.1016/j.infrared.2017.05.007. https://www.sciencedirect.com/science/article/pii/S1350449517300725
Arjovsky M, Chintala S, Bottou L (2017) Wasserstein Generative Adversarial Networks. In: Proceedings of the 34th International Conference on Machine Learning, PMLR pp. 214–223. https://proceedings.mlr.press/v70/arjovsky17a.html
Gulrajani I, Ahmed F, Arjovsky M, Dumoulin V, Courville AC (2017) Improved Training of Wasserstein GANs. In: Advances in Neural Information Processing Systems, vol. 30. Curran Associates, Inc.. https://proceedings.neurips.cc/paper/2017/hash/892c3b1c6dccd52936e27cbd0ff683d6-Abstract.html
Zhang L (2008) In situ image segmentation using the convexity of illumination distribution of the light sources. IEEE transactions on pattern analysis and machine intelligence 30(10):1786–1799
Shreyamsha Kumar BK (2015) Image fusion based on pixel significance using cross bilateral filter. Signal, Image and Video Processing 9(5):1193–1204. https://doi.org/10.1007/s11760-013-0556-9
Ma J, Chen C, Li C, Huang J (2016) Infrared and visible image fusion via gradient transfer and total variation minimization. Inf Fusion 31:100–109. https://doi.org/10.1016/j.inffus.2016.02.001. https://www.sciencedirect.com/science/article/pii/S156625351630001X
Bavirisetti DP, Xiao G, Liu G (2017) Multi-sensor image fusion based on fourth order partial differential equations. In: 2017 20th International Conference on Information Fusion (Fusion), pp. 1–9. https://doi.org/10.23919/ICIF.2017.8009719
Xu H, Ma J, Jiang J, Guo X, Ling H (2022) U2Fusion: A Unified Unsupervised Image Fusion Network. IEEE Trans. Pattern Anal Mach Intell 44(1):502–518. https://doi.org/10.1109/TPAMI.2020.3012548
Li H, Wu XJ, Kittler J (2020) MDLatLRR: A Novel Decomposition Method for Infrared and Visible Image Fusion. IEEE Trans. Image Process 29:4733–4746. https://doi.org/10.1109/TIP.2020.2975984
Ma J, Zhang H, Shao Z, Liang P, Xu H (2021) GANMcC: A Generative Adversarial Network With Multiclassification Constraints for Infrared and Visible Image Fusion. IEEE Trans Instrum Meas 70:1–14. https://doi.org/10.1109/TIM.2020.3038013
Meher B, Agrawal S, Panda R, Abraham A (2019) A survey on region based image fusion methods. Inf Fusion 48, 119–132 (2019-08-01). https://doi.org/10.1016/j.inffus.2018.07.010. https://www.sciencedirect.com/science/article/pii/S1566253517307583
Han Y, Cai Y, Cao Y, Xu X (2013) A new image fusion performance metric based on visual information fidelity. Inf Fusion 14(2):127–135. https://doi.org/10.1016/j.inffus.2011.08.002. https://www.sciencedirect.com/science/article/pii/S156625351100056X
Deshmukh M, Bhosale U, Deshmukh M, Engg SCO, Mumbai N, Bhosale U (2010) Image Fusion and Image Quality Assessment of Fused Images. International Journal of Image Processing (IJIP
Wang Z, Bovik A, Sheikh H (2004) Simoncelli E (2004) Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process. 13(4):600–612. https://doi.org/10.1109/TIP.2003.819861
Aslantas V (2015) Bendes E (2015) A new image quality metric for image fusion: The sum of the correlations of differences. AEU - International J Electron Commun 69(12):1890–1896. https://doi.org/10.1016/j.aeue.2015.09.004. https://www.sciencedirect.com/science/article/pii/S1434841115002691
Haghighat M, Razian MA (2014) Fast-fmi: non-reference image fusion metric.In: 2014 IEEE 8th International Conference on Application of Information and Communication Technologies (AICT), IEEE pp. 1–3
Zhang X, Ye P, Leung H, Gong K, Xiao G (2020) Object fusion tracking based on visible and infrared images: A comprehensive review. Inf Fusion 63:166–187. https://doi.org/10.1016/j.inffus.2020.05.002. https://www.sciencedirect.com/science/article/pii/S1566253520302657
Yang T, Xu P, Hu R, Chai H, Chan AB (2020) ROAM: Recurrently Optimizing Tracking Model. pp. 6718–6727. https://openaccess.thecvf.com/content_CVPR_2020/html/Yang_ROAM_Recurrently_Optimizing_Tracking_Model_CVPR_2020_paper.html
Acknowledgements
The authors gratefully acknowledge the financial supports by The National Science Foundation of China(62203224), Shanghai Special Plan for Local Colleges and Universities for Capacity Building (22010501300).
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Conflicts of interest
We declare that we have no financial and personal relationships with other people or organizations that can inappropriately influence our work, there is no professional or other personal interest of any nature or kind in any product, service or company that could be construed as influencing the position presented in the manuscript entitled "Manuscript".
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
About this article
Cite this article
Wang, X., Liu, G., Tang, L. et al. ICWGAN-GP: an image fusion method based on infrared compensator and wasserstein generative adversarial network with gradient penalty. Appl Intell 53, 27637–27654 (2023). https://doi.org/10.1007/s10489-023-04933-6
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10489-023-04933-6