Abstract
Infrared and visible image fusion combines the high resolution, rich structure of visible images with the remarkable information of infrared images for a wide range of applications in tasks such as target detection and segmentation. However, the representation and retention of significant and non-significant structures in fused images is still a big challenge. To address this issue, we proposed a novel infrared and visible image fusion approach based on the theory of fuzzy sets. First, we proposed a novel filter that integrates the SV-bitonic filter into a least squares model. By leveraging both global and local image features, this new filter achieved edge preservation and smoothness while effectively decomposing the image into structure and base layers. Moreover, for the fusion of the structure layers, according to the salient structural characteristics, we proposed a feature extraction method based on fuzzy inference system. Additionally, the intuitionistic fuzzy sets similarity measure was utilized to extract details from the residual structure layer. Extensive experiments demonstrate that the proposed model outperforms the state-of-the-art methods on publicly available datasets. The code is available at https://github.com/JEI981214/FUFusion_PRCV.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Zhou, H., Sun, M., Ren, X., Wang, X.Y.: Visible-thermal image object detection via the combination of illumination conditions and temperature information. Remote Sens. 13, 3656 (2021)
Chandrakanth, V., Murthy, V., Channappayya, S.S.: Siamese cross-domain tracker design for seamless tracking of targets in RGB and thermal videos. IEEE Trans. Artif. Intelli. 4, 161–172 (2022)
Wu, W., Qiu, Z.M., Zhao, M., Huang, Q.H., Lei, Y.: Visible and infrared image fusion using NSST and deep Boltzmann machine. Optik 157, 334–342 (2018)
Zhang, X., Demiris, Y.: Visible and infrared image fusion using deep learning. IEEE Trans. Pattern Anal. Mach. Intelli. 45, 10535–10554 (2023)
Gao, Y., Ma, S.W., Liu, J.J.: DCDR-GAN: a densely connected disentangled representation generative adversarial network for infrared and visible image fusion. IEEE Trans. Circ. Syst. Video Technol. 33, 549–561 (2023)
Zhao, Z.X., Xu, S., Zhang, J.S., Liang, C.Y., Zhang, C.X., Liu, J.M.: Efficient and model-based infrared and visible image fusion via algorithm unrolling. IEEE Trans. Circ. Syst. Video Technol. 32, 1186–1196 (2022)
Chang, Z.H., Feng, Z.X., Yang, S.Y., Gao, Q.W.: AFT: adaptive fusion transformer for visible and infrared images. IEEE Trans. Image Process. 32, 2077–2092 (2023)
Tang, L.F., Xiang, X.Y., Zhang, H., Gong, M.Q., Ma, J.Y.: DIVFusion: darkness-free infrared and visible image fusion. Inform. Fusion 91, 477–493 (2023)
Yang, Y., Zhang, Y., Huang, S., Zuo, Y., Sun, J.: Infrared and visible image fusion using visual saliency sparse representation and detail injection model. IEEE Trans. Instrum. Meas. 70, 1–15 (2021)
Li, X., Zhou, F., Tan, H.: Joint image fusion and denoising via three-layer decomposition and sparse representation. Know.-Based Syst. 224, 107087 (2021)
Li, H., Wu, X.J., Kittler, J.: MDLatLRR: a novel decomposition method for infrared and visible image fusion. IEEE Trans. Image Process. 29, 4733–4746 (2020)
Mo, Y., Kang, X.D., Duan, P.H., Sun, B., Li, S.T.: Attribute filter based infrared and visible image fusion. Inform. Fusion 75, 41–54 (2021)
Ren, L., Pan, Z., Cao, J., Zhang, H., Wang, H.: Infrared and visible image fusion based on edge-preserving guided filter and infrared feature decomposition. Signal Process. 186, 108108 (2021)
Treece, G.: Morphology-based noise reduction: structural variation and thresholding in the bitonic filter. IEEE Trans. Image Process. 186, 336–350 (2020)
Liu, W., Zhang, P.P., Chen, X.G., Shen, C.H., Huang, X.L., Yang, J.: Embedding bilateral filter in least squares for efficient edge-preserving image smoothing. IEEE Trans. Circuits Syst. Video Technol. 30, 23–35 (2020)
Alshennawy, A.A., Aly, A.A.: Edge detection in digital images using fuzzy logic technique. Int. J. Comput. Inform. Eng. 3, 540–548 (2009)
Li, S., Kang, X., Hu, J.: Image fusion with guided filtering. IEEE Trans. Image Process. 22, 2864–2875 (2013)
Liu, J.Y., Fan, X., Jiang, J., Liu, R.S., Luo, Z.X.: Learning a deep multi-scale feature ensemble and an edge-attention guidance for image fusion. IEEE Trans. Circuits Syst. Video Technol. 32, 105–119 (2022)
Zhou, H.B., Wu, W., Zhang, Y.D., Ma, J.Y., Ling, H.B.: Semantic-supervised infrared and visible image fusion via a dual-discriminator generative adversarial network. IEEE Trans. Multimedia 25, 635–648 (2023)
Tang, L.F., Yuan, J.T., Ma, J.Y.: Image fusion in the loop of high-level vision tasks: a semantic-aware real-time infrared and visible image fusion network. Inform. Fusion 82, 28–42 (2022)
Chen, J., Li, X.J., Luo, L.B., Mei, X.G., Ma, J.Y.: Infrared and visible image fusion based on target-enhanced multiscale transform decomposition. Inf. Sci. 508, 64–78 (2020)
Ma, J., Zhang, H., Shao, Z., Liang, P., Han, X.: GANMcC: A generative adversarial network with multiclassification constraints for infrared and visible image fusion. IEEE Trans. Instrum. Meas. 70, 1–14 (2021)
Li, H., Wu, X.-J.: Infrared and visible image fusion using Latent low-rank representation. arXiv preprint arXiv:1804.08992 (2022)
Qu, G., Zhang, D., Yan, P.: Information measure for performance of image fusion. Electron. lett. 38, 313–315 (2002)
Wang, Q., Shen, Y., Jin, J.: Performance evaluation of image fusion techniques. Image Fusion Algorithms Appl. 19, 469–492 (2008)
Xydeas, C.S., Petrovic, V.: Objective image fusion performance measure. Electron. Lett. 36, 308–309 (2000)
Wang, P.-W., Liu, B.: A novel image fusion metric based on multi-scale analysis. In: Proceedings of the 2008 9th International Conference on Signal Processing, pp. 965–968 (2008)
Wang, Z., Bovik, A.C., Sheikh, H.R., Simoncelli, E.P.: Image quality assessment: from error measurement to structural similarity. IEEE Trans. Image Process. 13, 600–613 (2004)
Han, Y., Cai, Y.Z., Cao, Y., Xu, X.M.: A new image fusion performance metric based on visual information fidelity. Inform. Fusion 14, 127–135 (2013)
Acknowledgements
This research was supported by the National Natural Science Foundation of China (Grant No. 62201149).
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2024 The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd.
About this paper
Cite this paper
Jie, Y., Chen, Y., Li, X., Yi, P., Tan, H., Cheng, X. (2024). FUFusion: Fuzzy Sets Theory for Infrared and Visible Image Fusion. In: Liu, Q., et al. Pattern Recognition and Computer Vision. PRCV 2023. Lecture Notes in Computer Science, vol 14426. Springer, Singapore. https://doi.org/10.1007/978-981-99-8432-9_37
Download citation
DOI: https://doi.org/10.1007/978-981-99-8432-9_37
Published:
Publisher Name: Springer, Singapore
Print ISBN: 978-981-99-8431-2
Online ISBN: 978-981-99-8432-9
eBook Packages: Computer ScienceComputer Science (R0)