Elsevier

Knowledge-Based Systems

Volume 258, 22 December 2022, 110041
Knowledge-Based Systems

Attention-guided dynamic multi-branch neural network for underwater image enhancement

https://doi.org/10.1016/j.knosys.2022.110041Get rights and content

Abstract

Underwater images often suffer from quality deteriorations such as color deviation, reduced contrast, and blurred details due to wavelength-dependent light absorption and scattering in water media. Recently, convolutional neural networks (CNNs) have achieved impressive success in underwater image enhancement (UIE). However, there is still room for improvement in terms of representational capability and receptive field (RF)/channel variant ability in almost all CNN-based UIE networks. To treat these problems, we propose an attention-guided dynamic multibranch neural network (ADMNNet) to obtain high-quality underwater images. Different from existing CNN-based UIE networks that generally share a fixed RF size of artificial neurons in one feature layer, we propose an attention-guided dynamic multibranch block (ADMB) to boost the diversity of feature representations by merging the properties of different RFs into a single-stream structure. Concretely, ADMB includes two main components, namely, a dynamic feature selection module (DFSM) and a multiscale channel attention module (MCAM). Inspired by the selective kernel mechanism in the visual cortex, we incorporate a nonlinear strategy into the DFSM that allows the neurons to adjust their RF sizes reasonably by using soft attention to achieve dynamic fusion of multiscale features. In the MCAM, channel attention is designed to exploit the interdependencies among the channelwise features extracted from different branches. Our ADMNNet can obtain better visual quality of underwater images captured under diverse scenarios and achieves superior qualitative and quantitative performance compared to state-of-the-art UIE methods.

Introduction

Underwater image enhancement (UIE) is a fundamental operation in the computer vision community that has become a hot spot in the field of image processing in recent years. The main purpose of UIE is to recover a clean image by eliminating degradations (e.g., color deviation and low contrast caused by wavelength-dependent attenuation) from its corresponding degraded version [1], [2], [3]. Research on this problem can be used in latent applications, such as underwater detection [4] and marine environmental surveillance [5]. However, it is an extremely challenging and ill-posed task, owing to medium attenuation properties or the diversity of underwater image distributions.

Studies show that earlier techniques expressly applied image priors handcrafted with empirical observations. These prior-based techniques often perform worse than expected because constructing strong image priors is difficult and usually fails to generalize. Recently, the rapid development of convolutional neural networks (CNNs) has promoted new strategies and perspectives for problem solving in the field of image processing. Therefore, CNNs have emerged as a dominant UIE approach, which can achieve state-of-the-art results with extraordinary feature representation capabilities.

Natural scenarios possess different areas, which should be taken into account at distinct scales. For instance, smooth areas correspond to features at larger scales, while the features in textured areas correspond to smaller ones [6]. However, most CNN-based UIE methods generally possess a finite receptive field (RF), which makes it difficult to estimate multiscale features. This inevitably limits their applicability to the diversity of water types and complex degradation levels.

To ameliorate the aforementioned problems, we present an effective CNN-based UIE solution. Following this intuition, to accomplish this challenging task, we build on two main facts/ observations. First, a few UIE techniques [7], [8] based on multiscale CNNs have been introduced, in which a linear manner is used to merge multiscale features. However, numerous studies [9], [10], [11] have demonstrated that the RF sizes of neurons in the same region (e.g., the primary visual cortex) are different but naturally adjusted via the stimulus. Therefore, these techniques cannot truly imitate the astonishing capability of neurons to adaptively integrate visual inputs, thereby restricting their performance. Second, enhancing underwater images is a wavelength-sensitive task due to variable levels of light attenuation for different wavelengths [12], [13], [14]. Each channelwise feature in an image embodies distinct types of input information. Some features may help to cope with the issues of color deviation, and others may contribute to improving the low contrast. If the interdependencies across channels are not taken into consideration, this will weaken the representational power of the network.

Based on the above motivation, a building block called the attention-guided dynamic multibranch block (ADMB) is introduced, which extracts the desired dynamic information from inputs at different branches. ADMB is mainly composed of a dynamic feature selection module (DFSM) and a multiscale channel attention-guided module (MCAM). To effectively advance the UIE task, multiple ADMB units are stacked in an end-to-end architecture, which is termed the attention-guided dynamic multibranch neural network (ADMNNet). Concretely, DFSM is proposed to exploit the gain of RF properties, which is responsible for receiving multiple branches with different RF sizes. To produce a final and global representation for the selection weights, the information from multiple branches is merged. An attention mechanism is applied to estimate selection weights, which can adaptively underline the most representative features. MCAM is proposed to improve the channel variant ability of the network. The proposed ADMNNet can effectively remove color deviation and improve detailed information, as shown in Fig. 1. Briefly, the significant contributions of this work are highlighted as follows.

We propose an attention-guided dynamic multibranch neural network (ADMNNet) to achieve superiority and adaptability for complex and numerous underwater images. Extensive experiments show that the proposed model introduces excellent robustness and flexibility when compared with state-of-the-art methods.

We develop a dynamic feature selection module (DFSM), in which the RF size of neurons can be adaptively modified by stimulation. More importantly, soft attention is used to fulfill the selective kernel mechanism between multiscale features. Put differently, our network has a self-adaptive adjustment function based on the contextual information of the input.

We design a multiscale channel attention-guided module (MCAM), which can be leveraged simultaneously to explore channelwise information through a more effective approach.

The rest of this paper is organized as follows: Section 2 introduces related work. The proposed method is presented in Section 3. Section 4 evaluates and compares the experimental results. Section 5 concludes the paper.

Section snippets

Related work

Handcrafted prior-based methods. Based on the underwater physical imaging model [17], a large number of handcrafted methods [18], [19], [20] execute the enhancement procedure in an inverse manner. Without any extra information, specific priors [21] as constraints are used to estimate the derived parameters (background light and transmission map) of the physical model [22]. For example, Drews et al. [23] proposed the underwater dark channel prior (UDCP), which compensated for the attenuation by

Proposed method

In this section, we describe an end-to-end attention-guided dynamic multibranch neural network (ADMNNet) for underwater image enhancement tasks. ADMNNet includes the core components of the proposed attention-guided dynamic multibranch block (ADMB): (a) dynamic feature selection module (DFSM) and (b) multiscale channel attention-guided module (MCAM). The overall architecture of our ADMNNet is illustrated in Fig. 2. Given an underwater image IR3×H×W as input, ADMNNet first applies a depthwise

Experimental settings

Datasets. To train ADMNNet, both real-world and synthetic underwater images are used. First, 800 paired images are randomly selected from the UIEB [34] dataset as the training set. In detail, the UIEB dataset contains 890 real underwater images with manually selected reference images. For testing, the remaining 90 images with references are employed, denoted as Test-R90 [34]. Although the original images in the UIEB dataset have different levels of contrast reduction and diverse scenes, the

Conclusion

In this paper, we propose an efficient yet simple attention-guided dynamic multibranch neural network (ADMNNet) for underwater image enhancement tasks. To learn the feature representations from different branches, an attention-guided dynamic multibranch block (ADMB) is designed. The ADMB, as the core component of our model, is composed of the dynamic feature selection module (DFSM) and multiscale channel attention-guided module (MCAM). Specifically, our DFSM implicitly models multiscale

CRediT authorship contribution statement

Xiaohong Yan: Writing – original draft, Writing – review & editing, Conceptualization, Methodology. Wenqiang Qin: Software. Yafei Wang: Writing – review & editing, Funding acquisition. Guangyuan Wang: Validation, Data curation. Xianping Fu: Supervision, Project administration, Funding acquisition.

Declaration of Competing Interest

The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.

Acknowledgment

The authors sincerely thank the editors and anonymous reviewers for the very helpful and kind comments to assist in improving the presentation of our paper. This work was supported in part by the National Natural Science Foundation of China under Grant 62176037, Grant 62002043, and Grant 61802043, by the Liaoning Revitalization Talents Program, China under Grant XLYC1908007, by the Foundation of Liaoning Key Research and Development Program, China under Grant 201801728, by the Dalian Science

References (57)

  • GangisettyS. et al.

    FloodNet: Underwater image restoration based on residual dense learning

    Signal Process., Image Commun.

    (2022)
  • LiH. et al.

    Haze transfer and feature aggregation network for real-world single image dehazing

    Knowl.-Based Syst.

    (2022)
  • YinS. et al.

    A novel image-dehazing network with a parallel attention block

    Pattern Recognit.

    (2020)
  • ZhangW. et al.

    Underwater image enhancement via minimal color loss and locally adaptive contrast enhancement

    IEEE Trans. Image Process.

    (2022)
  • MairalJ. et al.

    Learning multiscale sparse representations for image and video restoration

    Multiscale Model. Simul.

    (2008)
  • YangH.H. et al.

    LAFFNet: A lightweight adaptive feature fusion network for underwater image enhancement

  • LiuS. et al.

    Adaptive learning attention network for underwater image enhancement

    IEEE Robot. Autom. Lett.

    (2022)
  • HubelD.H. et al.

    Receptive fields, binocular interaction and functional architecture in the cat’s visual cortex

    J. Physiol.

    (1962)
  • LiuH. et al.

    Single image super-resolution using multi-scale deep encoder-decoder with phase congruency edge map guidance

    Inform. Sci.

    (2018)
  • GaoW. et al.

    Deep neural networks for sensor-based human activity recognition using selective kernel convolution

    IEEE Trans. Instrum. Meas.

    (2021)
  • MiZ. et al.

    Multi-purpose oriented real-world underwater image enhancement

    IEEE Access

    (2020)
  • YanX. et al.

    A natural-based fusion strategy for underwater image enhancement

    Multimed. Tools Appl.

    (2022)
  • AncutiC. et al.

    Enhancing underwater images and videos by fusion

  • LiC. et al.

    Underwater image enhancement via medium transmission-guided multi-color space embedding

    IEEE Trans. Image Process.

    (2021)
  • McglameryB.L.

    A computer model for underwater camera systems

    Ocean Opt. VI

    (1980)
  • XieJ. et al.

    A variational framework for underwater image dehazing and deblurring

    IEEE Trans. Circuits Syst. Video Technol.

    (2022)
  • DrewsJ.P. et al.

    Transmission estimation in underwater single images

  • HeK. et al.

    Single image haze removal using dark channel prior

    IEEE Trans. Pattern Anal. Mach. Intell.

    (2011)
  • Cited by (9)

    • Underwater Organism Color Fine-Tuning via Decomposition and Guidance

      2024, Proceedings of the AAAI Conference on Artificial Intelligence
    View all citing articles on Scopus
    View full text