Underwater image enhancement based on conditional generative adversarial network

https://doi.org/10.1016/j.image.2019.115723Get rights and content

Abstract

Underwater images play an essential role in acquiring and understanding underwater information. High-quality underwater images can guarantee the reliability of underwater intelligent systems. Unfortunately, underwater images are characterized by low contrast, color casts, blurring, low light, and uneven illumination, which severely affects the perception and processing of underwater information. To improve the quality of acquired underwater images, numerous methods have been proposed, particularly with the emergence of deep learning technologies. However, the performance of underwater image enhancement methods is still unsatisfactory due to lacking sufficient training data and effective network structures. In this paper, we solve this problem based on a conditional generative adversarial network (cGAN), where the clear underwater image is achieved by a multi-scale generator. Besides, we employ a dual discriminator to grab local and global semantic information, which enforces the generated results by the multi-scale generator realistic and natural. Experiments on real-world and synthetic underwater images demonstrate that the proposed method performs favorable against the state-of-the-art underwater image enhancement methods.

Introduction

Underwater images carry significant information of the underwater environment and are widely-used for exploring, recognizing, and monitoring underwater world. High-quality underwater images are desired in practical applications, which can guarantee the reliability of analysis and processing. However, suffering from the effects of light absorption and scattering, underwater images have obvious quality degradation issues, such as color distortion, low contrast, blurring, low light, uneven illumination, etc [1]. Concretely, the wavelength related attenuation causes color casts while the scattering brings a distance-dependent component into a clear image, which causes low-contrast of underwater images. Besides, underwater images also suffer from the noise and blurring induced by the cameras and motion. Several typical quality degraded underwater images are presented in Fig. 1. Such degraded underwater images limit their further applications in underwater object detection, underwater scene understanding, underwater robot inspection, underwater 3D reconstruction, to name a few [2], [3].

To obtain high-quality underwater images, the existing methods range from polarization filters, reversing physical models to deep learning-based technologies [4]. Due to the diversity of underwater image degradation [5], the performance of current methods has room to be improved. In recent years, much attention has been paid to deep learning-based visual tasks. Moreover, the deep learning-based methods have achieved encouraging results in many visual tasks [6], [7], [8], [9], [10], [11], [12]. Nevertheless, the deep learning-based underwater image enhancement methods usually fall behind the traditional underwater image enhancement methods. The main reasons focus on the limited and monotonous training data as well as ineffective network structures [1]. More recently, the success of cGANs [13] inspires us to explore the performance of cGANs in underwater image enhancement. Additionally, we also propose to solve the issues of limited underwater image training data based on a kind of random sampling strategy and the blending of synthetic and real data. The promising performance of recent low-level visual tasks [14], [15], [16], [17] also encourages our method. Experimental results on synthetic and real-world underwater image datasets demonstrate the robust and decent performance of our proposed method.

The contributions of this paper are summarized as follows.

  • We propose a conditional generative adversarial network (cGAN) for underwater image enhancement. To the best of our knowledge, this is the first cGAN with dual discriminator for underwater image enhancement problem.

  • To produce more realistic and natural results, we develop a dual discriminator which determines the reliability of generated results from different views.

  • Our proposed method can improve the contrast and remove color casts of underwater images, and achieves state-of-the-art performance on diverse scenes.

The remainder of this paper is scheduled as follows. In Section 2, we briefly review the related works, including single image dehazing methods, underwater image enhancement methods, and cGANs in low-level visual tasks. In Section 3, the proposed cGAN for underwater image enhancement is presented and analyzed. In Section 4, extensive experiments are conducted to demonstrate the effectiveness of the proposed method and the advantages over current methods. In Section 5, we discuss and conclude this paper.

Section snippets

Related work

Similar to underwater images, the images captured under hazy day also suffer from the effect of light scattering. Some underwater image enhancement methods borrow some techniques or ideas from the research area of single image dehazing. Thus, single image dehazing is related to our underwater image enhancement. Besides, as the key network skeleton, the cGANs are also related to our work. In this section, we will briefly review the related works including single image dehazing, underwater image

Our approach

In this section, we introduce the proposed approach, including the network structure and parameter settings (generator and discriminator), loss function, and implementation details.

Experiments

We compare the proposed cGN for underwater image enhancement with several state-of-the-art underwater image enhancement methods on both synthetic and real-world underwater images. The competing methods are UDCP [33], RED [35], UIBLA [38], and UWCNN [43]. The results of these methods are generated by the source codes provided by the authors with the recommended parameter settings. In [43], ten UWCNN models trained for different types of water are provided. We run these ten models and produce the

Conclusion

In this paper, we proposed a cGAN for underwater image enhancement. We use a multi-scale generator for extracting larger receptive fields while the proposed dual discriminator well pushes the generator towards better visual quality. Also, we propose two strategies to augment the training data, which is useful for solving the issues of limited training data in low-level visual tasks. Experimental results show that our method achieves visually pleasing and realistic results and surpasses the

Acknowledgments

This work was supported in part by the National Natural Science Foundation under Grant 61601194, 11573011. Natural Science Foundation of Jiangsu province under Grant BK20191469, the “Double creation talents” science and technology deputy general manager project in Jiangsu Province (2017), and in part by Lianyungang “521” project, Lianyungang “Haiyan” funded project in China, and the project of Collaborative Innovation Center of Advanced Ship and Marine Equipment/Marine Equipment and Technology

References (51)

  • AnwarS. et al.

    Class-specific image deblurring

  • GuoC. et al.

    Hierarchical features driven residual learning for depth map super-resolution

    IEEE Trans. Image Process.

    (2019)
  • WangW. et al.

    Attentive fashion grammar network for fashion landmark detection and clothing category classification

  • AnwarS. et al.

    Category-specific object image denoising

    IEEE Trans. Image Process.

    (2017)
  • WangW. et al.

    Revisiting video saliency: A large-scale benchmark and a new model

  • LiC. et al.

    Nested network with two-stream pyramid for salient object detection in optical remote sensing images

    IEEE Trans. Geosci. Remote Sens.

    (2019)
  • CongR. et al.

    An iterative co-saliency framework for RGBD images

    IEEE Trans. Cybernet.

    (2019)
  • MirzaM. et al.

    Conditional generative adversarial nets

    (2014)
  • ZhangH. et al.

    Image de-raining using a conditional generative adversarial network

    IEEE Trans. Circuits Syst. Video Technol.

    (2019)
  • CongR. et al.

    Co-saliency detection for RGBD images based on multi-constraint feature matching and cross label propagation

    IEEE Trans. Image Process.

    (2018)
  • CongR. et al.

    HSCS: Hierarchical sparsity based co-saliency detection for RGBD images

    IEEE Trans. Multimedia

    (2019)
  • HeK. et al.

    Single image haze removal using dark channel prior

    IEEE Trans. Pattern Anal. Mach. Intell.

    (2011)
  • MengG. et al.

    Efficient image dehazing with boundary constraint and contextual regularization

  • ZhuQ. et al.

    A fast single image haze removal algorithm using color attenuation prior

    IEEE Trans. Image Process.

    (2015)
  • BermanD. et al.

    Non-local image dehazing

  • Cited by (106)

    • Dual branch Transformer-CNN parametric filtering network for underwater image enhancement

      2024, Journal of Visual Communication and Image Representation
    View all citing articles on Scopus

    No author associated with this paper has disclosed any potential or pertinent conflicts which may be perceived to have impending conflict with this work. For full disclosure statements refer to https://doi.org/10.1016/j.image.2019.115723.

    View full text