Elsevier

Knowledge-Based Systems

Volume 213, 15 February 2021, 106617
Knowledge-Based Systems

A deep learning based image enhancement approach for autonomous driving at night

https://doi.org/10.1016/j.knosys.2020.106617Get rights and content

Abstract

Images of road scenes in low-light situations are lack of details which could increase crash risk of connected autonomous vehicles (CAVs). Therefore, an effective and efficient image enhancement model for low-light images is necessary for safe CAV driving. Though some efforts have been made, image enhancement still cannot be well addressed especially in extremely low light situations (e.g., in rural areas at night without street light). To address this problem, we developed a light enhancement net (LE-net) based on the convolutional neural network. Firstly, we proposed a generation pipeline to transform daytime images to low-light images, and then used them to construct image pairs for model development. Our proposed LE-net was then trained and validated on the generated low-light images. Finally, we examined the effectiveness of our LE-net in real night situations at various low-light levels. Results showed that our LE-net was superior to the compared models, both qualitatively and quantitatively.

Introduction

Benefiting from the rapid development of deep learning approaches in recent years, many computer vision applications have been developed for the design of advanced driver assistance systems (ADASs) and connected autonomous vehicles (CAVs) [1]. These applications mainly focus on object detection [2], [3], object classification [4], [5], object identification [6], [7], semantic segmentation [8], [9], [10], motion estimation [11], and surveillance systems [12]. However, most of the available computer vision applications are based on visible light cameras, and thus can only be used under the condition of normal light and clear weather [13], which makes most of the state-of-the-art models not suitable for nighttime images. Traffic safety statistics show that 51.1% of the U.S. fatal crashes happened in the nighttime (from 6 pm to 6 am), especially in rural areas with extremely low-illumination [14]. Hence, effectively enhancing nighttime images for a clear traffic environment is crucial for traffic safety [15], and should be incorporated into ADASs and CAVs for all-around-the-clock assistance.

However, the outline and appearance details of traffic participants are easily blurred at night, making it difficult to distinguish target objects from the background. Therefore, restoring the details of low-light images is a challenging task, especially for the rural low-light images. Various image enhancement methods have been proposed to solve this problem [16]. One classic method is histogram equalization (HE) [17]. HE makes brightness better distributed on the histogram, which can be used to enhance the local contrast without affecting the overall contrast. Another widely used method is gamma correction [18], [19], which increases the brightness of dark regions by compressing bright pixels. Besides HE and gamma correction, dark channel prior methods [20], retinex-based methods [21], [22], [23] and illumination map estimation methods [24] have also been proposed to address image enhancement problems, among which retinex-based methods are the most popular. More recently, driven by large datasets and the improvement of calculation capabilities, deep learning-based methods (e.g., CNN (convolutional neural network)) have shown great success in image enhancement applications [25], [26], [27], [28]. However, all these approaches only focused on the night situations with external light sources (e.g., urban roads with street light). There is a need for an image enhancement model that can accommodate darker scenarios as accidents are more likely to happen while driving in rural areas at night without street light [14].

To solve this problem and meet the requirements of high precision and real time for practical applications, we proposed a novel method for low-light image generation and an efficient end-to-end pipeline LE-net (light enhancement network) based on CNN to enhance images with different levels of illumination. Specifically, in our proposed low-light image generation method, two commonly used image transformation methods (i.e. gamma correction and contrast scale) were adopted to reduce the illumination of images. For further improvement, histogram matching was used to make the histogram distribution of generated images as close as possible to real night images, and a Gaussian mask based mixup strategy was used for local dimming because not all areas needed to be enhanced in an image. As for image enhancement, previous studies typically used large convolutional networks [25], [26], [27], [28]. A limitation of the applications with large convolutional networks is that they are computationally intensive and cannot meet the real time requirement for practical applications. In our proposed LE-net, a new efficient convolution was introduced to reduce the computation burden, and advanced technologies including linear bottleneck and inversed residual [29], self-attention distillation [30], and feature pyramid module [31] were used to improve the network performance. The main contributions of this study can be summarized as follows.

(1) We developed an effective and efficient CNN based network LE-net to enhance low-light images. The strong nonlinear mapping capability of our LE-net comprehensively models the relationship between our generated image pairs in various illumination situations so that our image enhancement model has a better generalization performance.

(2) We proposed a novel pipeline method for low-light image generation. This method can improve the robustness of our proposed image enhancement model by providing superficially realistic and naturalistic images in various illumination situations for training.

Our efforts can be adopted in the development of ADASs and CAVs to help avoid crashes at night especially in rural areas, which would significantly improve both driving safety for drivers and traffic safety for vulnerable traffic participants.

The remainder of this paper is organized into four sections. Firstly, Section 2 provides a review on the related methods including retinex-based methods and learning-based methods. Section 3 describes our proposed method to generate and enhance low-light images. The corresponding results and the superiority of our proposed models are presented and discussed in Section 4. Finally, Section 5 presents the main conclusions.

Section snippets

Retinex-Based methods

The retinex theory assumes that the observed image S is determined by the reflected light L and reflectivity R of an object, therefore the observed image S(x,y) is defined as: S(x,y)=Rx,yL(x,y)where (x,y) is the pixel coordinate, S(x,y) is the observed image, Rx,y is the reflection image (desired recovery), and L(x,y) is the illumination image.

Single-scale retinex (SSR) is one of the models developed based on retinex [32]. Mathematically, it works as follows:

Step1: Using logarithmic function

Methodology

To solve the above-mentioned image enhancement problems, we developed a CNN based pipeline LE-net to restore details of images which could help drivers and CAVs see in the dark. The pipeline includes three steps: (1) generating training pairs, (2) training the CNN based model, and (3) enhancing both generated low-light images and real night images.

Experiment results and discussions

We conducted experiments to examine the effectiveness of our data generation pipeline and dark image enhancement method. Fig. 7 illustrates the organization of this section for a clear overview. The superiority of our generation method for low-light images is described in Section 4.1. The superiority of our proposed LE-net is presented in Section 4.2 and Section 4.3 with respect to the qualitative and quantitative comparisons, respectively. Section 4.4 further examines the effectiveness of our

Conclusion

Restoring low-light images is a challenging task and is important for the design of ADASs and CAVs. As there are no paired images in BDD100K to train our proposed deep learning network, we propose a new data generation pipeline to generate low-light paired images. The generated images are close to reality, indicating that these images can be used as the training pairs of our proposed LE-net. The effectiveness of our LE-net is validated by comparing with other traditional and deep learning

CRediT authorship contribution statement

Guofa Li: Conceptualization, Methodology, Writing - original draft, Writing - review & editing, Funding acquisition. Yifan Yang: Conceptualization, Methodology, Validation, Formal analysis, Data curation, Visualization, Writing - original draft, Writing - review & editing. Xingda Qu: Writing - review & editing, Supervision. Dongpu Cao: Conceptualization. Keqiang Li: Resources, Supervision.

Declaration of Competing Interest

The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.

Acknowledgments

This study is supported by the National Natural Science Foundation of China (grant number: 51805332), the Natural Science Foundation of Guangdong Province, China (grant number: 2018A030310532), the Shenzhen Fundamental Research Fund, China (grant number: JCYJ20190808143415801 and JCYJ20190808142613246), and the Young Elite Scientists Sponsorship Program funded by the China Society of Automotive Engineers .

References (68)

  • X. Sun, L. Zheng, Dissecting person re-identification from the viewpoint of viewpoint, in: Proceedings of the IEEE...
  • DingY. et al.

    Adaptive exploration for unsupervised person re-identification

    ACM Trans. Multimed. Comput. Commun. Appl.

    (2020)
  • HeW. et al.

    A survey of human-centered intelligent robots: issues and challenges

    IEEE/CAA J. Autom. Sinica

    (2017)
  • LiJ. et al.

    An end-to-end load balancer based on deep learning for vehicular network traffic control

    IEEE Internet Things J.

    (2019)
  • ChenB.-H. et al.

    Error-optimized sparse representation for single image rain removal

    IEEE Trans. Ind. Electron.

    (2017)
  • NHTSA, Traffic Safety Facts 2017 (DOT HS 812 806), National Highway Traffic Safety Administration, U.S. Department of...
  • LiuS. et al.

    L0 Sparse regularization-based image blind deblurring approach for solid waste image restoration

    IEEE Trans. Ind. Electron.

    (2019)
  • ChenS.-D. et al.

    Contrast enhancement using recursive mean-separate histogram equalization for scalable brightness preservation

    IEEE Trans. Consum. Electron.

    (2003)
  • ZhengL. et al.

    Infrared traffic image enhancement algorithm based on dark channel prior and gamma correction

    Modern Phys. Lett. B

    (2017)
  • ChangY. et al.

    Automatic contrast-limited adaptive histogram equalization with dual gamma correction

    IEEE Access

    (2018)
  • PanJ. et al.

    Deblurring images via dark channel prior

    IEEE Trans. Pattern Anal. Mach. Intell.

    (2018)
  • LiM. et al.

    Structure-revealing low-light image enhancement via robust retinex model

    IEEE Trans. Image Process.

    (2018)
  • GaoY. et al.

    Naturalness preserved nonuniform illumination estimation for image enhancement based on retinex

    IEEE Trans. Multimed.

    (2018)
  • ShenJ. et al.

    Nighttime driving safety improvement via image enhancement for driver face detection

    IEEE Access

    (2018)
  • GuoX. et al.

    LIME: Low-light image enhancement via illumination map estimation

    IEEE Trans. Image Process.

    (2017)
  • ShenL. et al.

    MSR-Net: low-light image enhancement using deep convolutional network

    (2017)
  • RenW. et al.

    Low-light image enhancement via a deep hybrid network

    IEEE Trans. Image Process.

    (2019)
  • GuoY. et al.

    A pipeline neural network for low-light image enhancement

    IEEE Access

    (2019)
  • M. Sandler, A. Howard, M. Zhu, A. Zhmoginov, L.-C. Chen, MobileNetV2: inverted residuals and linear bottlenecks, in:...
  • Y. Hou, Z. Ma, C. Liu, C.C. Loy, Learning lightweight lane detection CNNs by self attention distillation, in:...
  • T.-Y. Lin, P. Dollar, R. Girshick, K. He, B. Hariharan, S. Belongie, Feature pyramid networks for object detection, in:...
  • JobsonD.J. et al.

    A multiscale retinex for bridging the gap between color images and the human observation of scenes

    IEEE Trans. Image Process.

    (1997)
  • JobsonD.J. et al.

    Properties and performance of a center/surround retinex

    IEEE Trans. Image Process.

    (1997)
  • YangX. et al.

    Implementing real-time RCF-retinex image enhancement method using CUDA

    J. Real-Time Image Process.

    (2019)
  • Cited by (136)

    • Edge-guided oceanic scene element detection

      2024, Knowledge-Based Systems
    View all citing articles on Scopus
    View full text