Skip to main content
Log in

LCDA-Net: Efficient Image Dehazing with Contrast-Regularized and Dilated Attention

  • Published:
Neural Processing Letters Aims and scope Submit manuscript

Abstract

To address the issues of incomplete dehazing and low dehazing efficiency in existing dehazing networks, this study introduces a Lightweight Contrast-Regularized Dilated Attention Network (LCDA-Net) for single-image dehazing. Initially, Attention Context Encoding (ACE) is employed to decompose the input image into high-frequency and low-frequency features. For the low-frequency features, which are significantly impacted by haze, a pyramid dehazing module based on large-kernel dilated convolutional attention is devised, facilitating efficient dehazing through complementary semantic information. In contrast, for high-frequency features, a detail enhancement module based on deformable convolution is designed to restore fine texture information. Subsequently, high-frequency and low-frequency features are merged to reconstruct a clear image. Lastly, a loss function is designed by incorporating contrast regularization and edge loss strategies, effectively guiding the network to generate more realistic images. In this network, depthwise separable convolutions replace traditional convolutions, significantly reducing model complexity while maintaining satisfactory dehazing performance. Experimental results on the RESIDE benchmark dataset demonstrate that, compared to other advanced methods, the proposed approach achieves superior dehazing outcomes for both synthetic and real haze images, effectively mitigating artifacts, distortions, and incomplete dehazing. The PSNR on the SOTS indoor and outdoor test sets reaches 31.73 dB and 29.31 dB, respectively, with a network parameter size of merely 2 M. Additionally, the proposed method exhibits the lowest model complexity while achieving optimal performance metrics and the highest FPS, indicating both its superior dehazing performance and low complexity.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6
Fig. 7
Fig. 8
Fig. 9
Fig. 10
Fig. 11
Fig. 12
Fig. 13
Fig. 14

Similar content being viewed by others

References

  1. Kumar A, Srivastava S (2020) Object detection system based on convolution neural networks using single shot multi-box detector. Proc Comput Sci 171:2610–2617

    Article  Google Scholar 

  2. Mo Y, Wu Y, Yang X, Liu F, Liao Y (2022) Review the state-of-the-art technologies of semantic segmentation based on deep learning. Neurocomputing 493:626–646

    Article  Google Scholar 

  3. Lauer J, Zhou M, Ye S, Menegas W, Schneider S, Nath T, Rahman MM, Di Santo V, Soberanes D, Feng G et al (2022) Multi-animal pose estimation, identification and tracking with deeplabcut. Nat Methods 19(4):496–504

    Article  Google Scholar 

  4. McCartney EJ (1976) Optics of the atmosphere: scattering by molecules and particles. New York

  5. He K, Sun J, Tang X (2010) Single image haze removal using dark channel prior. IEEE Trans Pattern Anal Mach Intell 33(12):2341–2353

    Google Scholar 

  6. Zhu Q, Mai J, Shao L (2015) A fast single image haze removal algorithm using color attenuation prior. IEEE Trans Image Process 24(11):3522–3533

    Article  MathSciNet  MATH  Google Scholar 

  7. Fattal R (2014) Dehazing using color-lines. ACM Trans Graph 34(1):1–14

    Article  Google Scholar 

  8. Berman D, Avidan S, et al (2016) Non-local image dehazing. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 1674–1682

  9. Liu J, Liu W, Sun J, Zeng T (2021) Rank-one prior: toward real-time scene recovery. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 14802–14810

  10. Cai B, Xu X, Jia K, Qing C, Tao D (2016) Dehazenet: an end-to-end system for single image haze removal. IEEE Trans Image Process 25(11):5187–5198

    Article  MathSciNet  MATH  Google Scholar 

  11. Ren W, Liu S, Zhang H, Pan J, Cao X, Yang M-H (2016) Single image dehazing via multi-scale convolutional neural networks. In: Proceedings of 14th European conference computer vision–ECCV 2016: , Amsterdam, The Netherlands, October 11-14, 2016, Part II 14. Springer, pp 154–169

  12. Li B, Peng X, Wang Z, Xu J, Feng D (2017) Aod-net: all-in-one dehazing network. In: Proceedings of the IEEE international conference on computer vision, pp 4770–4778

  13. Liu X, Ma Y, Shi Z, Chen J (2019) Griddehazenet: attention-based multi-scale network for image dehazing. In: Proceedings of the IEEE/CVF international conference on computer vision, pp 7314–7323

  14. Chen D, He M, Fan Q, Liao J, Zhang L, Hou D, Yuan L, Hua G (2019) Gated context aggregation network for image dehazing and deraining. In: 2019 IEEE winter conference on applications of computer vision (WACV), pp 1375–1383. IEEE

  15. Qin X, Wang Z, Bai Y, Xie X, Jia H (2020) Ffa-net: feature fusion attention network for single image dehazing. In: Proceedings of the AAAI conference on artificial intelligence, vol 34, pp 11908–11915

  16. Zhang X, Wang T, Wang J, Tang G, Zhao L (2020) Pyramid channel-based feature attention network for image dehazing. Comput Vis Image Underst 197:103003

    Article  Google Scholar 

  17. Song Y, He Z, Qian H, Du X (2023) Vision transformers for single image dehazing. IEEE Trans Image Process 32:1927–1941

    Article  Google Scholar 

  18. Yang Y, Wang C, Liu R, Zhang L, Guo X, Tao D (2022) Self-augmented unpaired image dehazing via density and depth decomposition. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 2037–2046

  19. Ullah H, Muhammad K, Irfan M, Anwar S, Sajjad M, Imran AS, Albuquerque VHC (2021) Light-dehazenet: a novel lightweight CNN architecture for single image dehazing. IEEE Trans Image Process 30:8968–8982

    Article  Google Scholar 

  20. Niu Z, Zhong G, Yu H (2021) A review on the attention mechanism of deep learning. Neurocomputing 452:48–62

    Article  Google Scholar 

  21. Su YZ, He C, Cui ZG, Li AH, Wang N (2023) Physical model and image translation fused network for single-image dehazing. Pattern Recogn 142:109700. https://doi.org/10.1016/j.patcog.2023.109700

    Article  Google Scholar 

  22. Wang N, Cui Z, Su Y, He C, Li A (2021) Multiscale supervision-guided context aggregation network for single image dehazing. IEEE Signal Process Lett 29:70–74

    Article  Google Scholar 

  23. Wu H, Qu Y, Lin S, Zhou J, Qiao R, Zhang Z, Xie Y, Ma L (2021) Contrastive learning for compact single image dehazing. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 10551–10560

  24. Vaswani A, Shazeer N, Parmar N, Uszkoreit J, Jones L, Gomez AN, Kaiser Ł, Polosukhin I (2017) Attention is all you need. Adv Neural Inf Process Syst 30

  25. Dosovitskiy A, Beyer L, Kolesnikov A, Weissenborn D, Zhai X, Unterthiner T, Dehghani M, Minderer M, Heigold G, Gelly S, et al (2020) An image is worth 16x16 words: transformers for image recognition at scale. arXiv preprint arXiv:2010.11929

  26. Lindeberg T (1994) Scale-space theory: a basic tool for analyzing structures at different scales. J Appl Stat 21(1–2):225–270

    Article  Google Scholar 

  27. Chen Y, Fan H, Xu B, Yan Z, Kalantidis Y, Rohrbach M, Yan S, Feng J (2019) Drop an octave: Reducing spatial redundancy in convolutional neural networks with octave convolution. In: Proceedings of the IEEE/CVF international conference on computer vision, pp 3435–3444

  28. Xu K, Yang X, Yin B, Lau RW (2020) Learning to restore low-light images via decomposition-and-enhancement. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 2281–2290

  29. Howard AG, Zhu M, Chen B, Kalenichenko D, Wang W, Weyand T, Andreetto M, Adam H (2017) Mobilenets: efficient convolutional neural networks for mobile vision applications. arXiv preprint arXiv:1704.04861

  30. Sandler M, Howard A, Zhu M, Zhmoginov A, Chen L-C (2018) Mobilenetv2: Inverted residuals and linear bottlenecks. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 4510–4520

  31. Zhang X, Zhou X, Lin M, Sun J (2018) Shufflenet: An extremely efficient convolutional neural network for mobile devices. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 6848–6856

  32. Ma N, Zhang X, Zheng H-T, Sun J (2018) Shufflenet v2: practical guidelines for efficient cnn architecture design. In: Proceedings of the European conference on computer vision (ECCV), pp 116–131

  33. Ioffe S, Szegedy C (2015) Batch normalization: accelerating deep network training by reducing internal covariate shift. In: International conference on machine learning, pp 448–456. PMLR

  34. Nie D, Lan R, Wang L, Ren X (2022) Pyramid architecture for multi-scale processing in point cloud segmentation. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 17284–17294

  35. Dai J, Qi H, Xiong Y, Li Y, Zhang G, Hu H, Wei Y (2017) Deformable convolutional networks. In: Proceedings of the IEEE international conference on computer vision, pp 764–773

  36. Shao Y, Li L, Ren W, Gao C, Sang N (2020) Domain adaptation for image dehazing. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 2808–2817

  37. You Y, Chen T, Sui Y, Chen T, Wang Z, Shen Y (2020) Graph contrastive learning with augmentations. Adv Neural Inf Process Syst 33:5812–5823

    Google Scholar 

  38. Tian Y, Sun C, Poole B, Krishnan D, Schmid C, Isola P (2020) What makes for good views for contrastive learning? Adv Neural Inf Process Syst 33:6827–6839

    Google Scholar 

  39. Chen T, Kornblith S, Norouzi M, Hinton G (2020) A simple framework for contrastive learning of visual representations. In: International conference on machine learning, pp 1597–1607. PMLR

  40. Simonyan K, Zisserman A (2014) Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:1409.1556

  41. Qu Y, Chen Y, Huang J, Xie Y (2019) Enhanced pix2pix dehazing network. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 8160–8168

  42. Liu S, Ding W, Liu C, Liu Y, Wang Y, Li H (2018) Ern: Edge loss reinforced semantic segmentation network for remote sensing images. Remote Sensing 10(9):1339

    Article  Google Scholar 

  43. Wang Z, Bovik AC, Sheikh HR, Simoncelli EP (2004) Image quality assessment: from error visibility to structural similarity. IEEE Trans Image Process 13(4):600–612

    Article  Google Scholar 

  44. Venkatanath N, Praneeth D, Bh MC, Channappayya SS, Medasani SS (2015) Blind image quality evaluation using perception based features. In: 2015 twenty first national conference on communications (NCC), pp 1–6. IEEE

  45. Mittal A, Soundararajan R, Bovik AC (2012) Making a “completely blind’’ image quality analyzer. IEEE Signal Process Lett 20(3):209–212

    Article  Google Scholar 

  46. Mittal A, Moorthy AK, Bovik AC (2012) No-reference image quality assessment in the spatial domain. IEEE Trans Image Process 21(12):4695–4708

    Article  MathSciNet  MATH  Google Scholar 

  47. Ren W, Ma L, Zhang J, Pan J, Cao X, Liu W, Yang M-H (2018) Gated fusion network for single image dehazing. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 3253–3261

  48. Zhang H, Patel VM (2018) Densely connected pyramid dehazing network. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 3194–3203

  49. Zhang J, Tao D (2019) Famed-net: a fast and accurate multi-scale end-to-end dehazing network. IEEE Trans Image Process 29:72–84

    Article  MathSciNet  MATH  Google Scholar 

  50. Li W, Fan G, Gan M (2023) Progressive encoding-decoding image dehazing network. Multimed Tools Appl, pp 1–23

Download references

Acknowledgements

This work was Supported by National Natural Science Foundation of China (Grant No. 12071126).

Author information

Authors and Affiliations

Authors

Contributions

XL and SC designed and carried out the experimental studies. SC was responsible for data analysis and interpretation. The main manuscript text and literature review were written by SC and ZW. All tables and figures were prepared by ZW and YC, who also participated in the design of the experiment. XL provided crucial equipment and research materials. All authors collectively reviewed and revised the manuscript and agreed to the final draft of the paper.

Corresponding author

Correspondence to Xun Luo.

Ethics declarations

Conflict of interest

The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Luo, X., Cao, S., Wang, Z. et al. LCDA-Net: Efficient Image Dehazing with Contrast-Regularized and Dilated Attention. Neural Process Lett 55, 11467–11488 (2023). https://doi.org/10.1007/s11063-023-11384-0

Download citation

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s11063-023-11384-0

Keywords

Navigation