Abstract
Images affected by haze usually present faded colours and loss of contrast, hindering the precision of methods devised for clear images. For this reason, image dehazing is a crucial pre-processing step for applications such as self-driving vehicles or tracking. Some of the most successful dehazing methods in the literature do not follow any physical model and are just based on either image enhancement or image fusion. In this paper, we present a procedure to allow these methods to accomplish the Koschmieder physical model, i.e., to force them to have a unique transmission for all the channels, instead of the per-channel transmission they obtain. Our method is based on coupling the results obtained for each of the three colour channels. It improves the results of the original methods both quantitatively using image metrics, and subjectively via a psychophysical test. It especially helps in terms of avoiding over-saturation and reducing colour artefacts, which are the most common complications faced by image dehazing methods.
GF and JVC have received funding from the British Government’s EPSRC programme under grant agreement EP/M001768/1. MB and JVC have received funding from the European Union’s Horizon 2020 research and innovation programme under grant agreement number 761544 (project HDR4EU) and under grant agreement number 780470 (project SAUCE), and by the Spanish government and FEDER Fund, grant ref. TIN2015-71537-P (MINECO/FEDER,UE).
You have full access to this open access chapter, Download conference paper PDF
Similar content being viewed by others
Keywords
1 Introduction
Images acquired in outdoor scenarios often suffer from the effects of atmospheric phenomena such as fog or haze. The main characteristic of these phenomena is light scatter. The scattering effect distorts contrast and colour in the image, decreasing the visibility of content in the scene and reducing the visual quality.
Koschmieder [15] defined a model of how the atmospheric phenomena affects the output images. The model depends on two parameters: a depth-dependent transmission (\(\varvec{t}\)), and the colour of the airlight (\(\varvec{A}\)). Mathematically, the model is written as
Here x is a particular image pixel, \(\varvec{J_{x,\cdot }}\) is the 1-by-3 vector of the R,G,B values at pixel x of the clear image (i.e., how the image would look without atmospheric scatter) and \(\varvec{I}_{x,\cdot }\) is the 1-by-3 vector of the R,G,B values at pixel x of the image presenting the scattering effect. We remark that the transmission \(\varvec{t}\) only depends on the depth of the image, and therefore it is supposed to be equal for the three colour channels.
Image dehazing methods -i.e. methods that given a hazy image \(\varvec{I}\), obtain a clear image \(\varvec{J}\)- are becoming crucial for computer vision, because there are several methods -for recognition and classification among other tasks- that are supposed to work in the wild. Some examples are those used for surveillance through CCTV cameras, tracking, or the self-driving of vehicles and drones. However, the vast majority of these methods are devised for clear images, and tend to fail under adverse weather conditions. Image dehazing methods can be roughly divided in two categories: (i) physical-based methods that estimate the transmission of the image and solve for the clear image by inverting Eq. 1 [3, 4, 7, 13, 18, 21, 23, 24, 27], and (ii) image processing methods that directly process the hazy image so as to obtain a dehazed image but without considering the previous equation (from now on, we will call these methods non-physical dehazing methods) [2, 6, 10,11,12, 25, 26].
In this paper we focus on non-physical dehazing methods. This type of methods are able to obtain state-of-the-art results, but may sometimes present over-saturated colours and colour artefacts mostly because a different transmission is obtained for each colour channel. An example of the problems just mentioned is shown in Fig. 1 where, from left to right, we show two original images, the results from the methods of Galdran et al. [11] (top) and Choi et al. [6] (bottom), and the results obtained by using the approach of this paper.
There are very few proven methods that specifically look at reducing the colour artefacts that appear in dehazed images. Matlin and Milanfar [17] proposed an iterative regression method to simultaneously perform denoising and dehazing. Li et al. [16] decomposed the image into high and low frequencies, performing the dehazing only in the low frequencies, thus avoiding blocking artifacts. Chen et al. [5] applied both a smoothing filter for the refinement of the transmission and an energy minimisation in the recovery phase to avoid the appearance of gradients in the output image that were not presented in the original image.
In this paper we present a post-processing model for non-physical dehazing methods that aims at providing an output image that accomplishes the physical constraints given by Eq. 1. Our method is based on a channel-coupling approach, and it is devised to obtain a single transmission for all the different colour channels. Furthermore, our method also improves on the estimation of the airlight colour.
2 Imposing a Physically Plausible Dehazing
In this section, we define our approach for the post-processing of non-physical dehazing methods. Our main goal is, given an original hazy image and the solution of a non-physical dehazing method, to obtain a single transmission and an airlight that minimise the error of Eq. 1. We can write this minimisation in matrix form as:
where, \(\varvec{1}\) is a N-by-1 vector that has a value of 1 in every entry, \(\varvec{t^*}\) is a N-by-1 vector that represents the transmission, \(\varvec{A^*}\) is a 1-by-3 vector that provides us with the airlight, \(\varvec{I}\), \(\varvec{J}\) are N-by-3 matrices representing the input image, and the non-physical dehazing solution, N is the number of pixels, \(\varvec{T^*}\) is a N-by-3 matrix consisting on the replication of \(\varvec{t^*}\) three times, and \(\odot \) represents the element-wise multiplication.
It is clear that to solve for this equation, we need to select an input guessing for either \(\varvec{A^{our}}\) or \(\varvec{t^{our}}\). This is not a problem, since a standard hypothesis used in many image dehazing works is to select \(\varvec{A^{our}} =[1,1,1]\). Equation 2 also teaches us that we should perform the minimisation iteratively in two different dimensions. When we look for \(\varvec{t^{our}}\) we should perform the minimisation for each pixel x of the image over the three colour channels, while when we look for \(\varvec{A^{our}}\) we should perform the minimisation for each colour channel c over all the pixels.
We now detail our iterative minimisation. Let us start by having \(\varvec{I}\), \(\varvec{J}\), and the initial guessing for \(\varvec{A^{our}}\). In this case we can solve for the value of \(\varvec{t^{our}}\) at each pixel value x using a least squares minimisation:
As stated in the introduction, \(\varvec{J_{x,\cdot }}\) and \(\varvec{I_{x,\cdot }}\) are the 1-by-3 colour vectors at pixel x. This least squares minimisation has the following solution
where \(^T\) denotes the transpose of the vector.
Once we have found the transmission value \(\varvec{t^{our}}\), we can refine the value of \(\varvec{A^{our}}\) via a least squares approach. In this case, as stated above, we perform the least squares minimisation over the pixels of the image for each of the three colour channels. Mathematically,
In this case \(\varvec{J_{\cdot ,c}}\) and \(\varvec{I_{\cdot ,c}}\) are N-by-1 vectors representing each different colour channel of the images -i.e. \(c=\{R,G,B\}\)-, N is the number of pixels, and \(\varvec{1}\) is also a N-by-1 vector that has 1 at every entry.
This minimisation leads to
where \(^T\) denotes the transpose of the vector.
Once this new \(\varvec{A^{our}}\) is obtained, we can keep the iterative approach going by further refining the previous \(\varvec{t^{our}}\) following again Eq. 3.
Finally, once the desired number of iterations are performed, and given \(\varvec{t^{our}}\), \(\varvec{A^{our}}\), and the original hazy image \(\varvec{I}\), we can obtain our output image \(\varvec{J^{our}}\) by solving for Eq. 1:
We want the reader to raise attention to the relation of this approach to the Alternative Least Squares (ALS) method introduced by Finlayson et al. [8]. As in the ALS method, we are following an iterative procedure for the minimisation of a norm-based function.
3 Experiments and Results
This section is divided into three parts. First, we show qualitative results for our approach when applied to different non-physical dehazing methods. This is followed by a quantitative analysis of our post-processing. The section ends with a subjective evaluation using a preference test. In all our results we have allowed our approach to perform 5 iterations, as we have found experimentally that they are enough to obtain stable results. We have initialised the iterative approach by supposing \(\varvec{A^{our}}=[1,1,1]\).
3.1 Qualitative Evaluation
In all the following figures, we show on the left the original hazy image, on the center the result of the selected dehazing method, and on the right the result obtained by our method.
Figure 2 shows the results for the EVID method [11]. We can see that the original method is inducing an odd increase of contrast in the nearby objects of the image, therefore provoking these objects to look unnatural (e.g. the nearby plants in the top image, and the gravestones in the bottom image). These problems are clearly alleviated in our results.
Figure 3 shows the results for the FVID method [12]. The biggest problem of this image dehazing method is the appearance of artefacts (located in the base of the bushes in the top image and in the sky in the bottom image). Also, the top image is clearly presenting an excessive unnatural contrast. All these problems are suppressed by our proposed approach.
Figure 4 shows the results for the DEFADE method [6]. This method over-enhances the colours, as it can be clearly seen in the green of the plants in the top image, and in the orange hue of the boy’s jacket in the bottom one. Once again, these problems are solved after applying our proposed post-processing.
Finally, Fig. 5 shows the results for the method of Wang et al. [26]. In this particular case, images present an unreasonable contrast. This fact provokes the appearance of unrealistic edges and colours (focus on the green of the grass and the closer bushes in the top image, and on the wall of the nearby building in the bottom image). Once again, these problems are mitigated once our method is applied.
3.2 Quantitative Evaluation
For this subsection, we have selected six standard hazy images that appear in most of the works dealing with image dehazing. They are shown in Fig. 6. Regarding the non-physical dehazing methods to be evaluated, we have selected the following five: the FVID [12], the DEFADE [6], the method of Wang et al. [26], and the use of the DehRet method by [10], considering as Retinex the variational approach of SCIE [9] and the Multiscale Retinex (MSCR) method. [22].
We have computed two different image quality metrics in order to evaluate our results: the Naturalness Image Quality Evaluator (NIQE) [20], and the BRISQUE metric [19]. We have selected these metrics as we do not have access to corresponding ground-truth (fog-free) images. Let us note that in the case there are ground-truth images available further metrics can also be considered [14].
NIQE is an error metric that states how natural an image is (the smaller the number, the higher the naturalness). Table 1 presents the mean and RMS results for this metric. Our method improves in all the cases except for the FVID method. In this last case, the mean for the original dehazing method and the mean for our approach is the same, and the RMS for our approach is slightly worse than the one for the original dehazing.
BRISQUE is a distortion-based metric that also tries to predict if an image looks natural based on scene statistics (the smaller the value, the better the result). Table 2 presents the results for this metric. In this case, our method outperforms all the others for all the cases.
3.3 Preference Test
We have also performed a preference test with the same set of images used in the previous subsection. In total, 7 observers completed the experiment. All observers were tested for normal colour vision. The experiment was conducted on a NEC SpectraView reference 271 monitor set to ‘sRGB’ mode. The display was viewed at a distance of approximately 70 cm so that 40 pixels subtended \(1^{\circ }\) of visual angle. Stimuli were generated running MATLAB (MathWorks) with functions from the Psychtoolbox. The experiment was conducted in a dark room.
Subjects were presented with three images: in the center the original hazy image, and at each side the result of the original dehazing method and the result of our post-processing approach. Let us note that the side for these two images was selected randomly, and therefore varied at each presentation. Subjects were asked to select the preferred dehazed image. The total number of comparisons was 30.
Results have been obtained following the Thurstone Case V Law of Comparative Judgement. Figure 7 shows the results for the whole set of comparisons (i.e., considering the 5 original dehazing methods together). We can clearly see that our method statistically outperforms the original dehazing methods.
A more detailed analysis that looks individually at each dehazing method is presented in Fig. 8. We can clearly see that our method greatly outperforms the results of the DEFADE, the Wang et al., and the DehRet-MSCR methods. In the case of the FVID and the DehRet-SRIE methods, our method is statistically equivalent to the original method. Let us note that these results are well aligned with those obtained on the previous subsection, as the two methods that are statistically equivalent to our post-processing were also the two methods for which our improvement in the metrics was smaller.
The results shown lead us to conclude that our method is very reliable, both quantitatively and subjectively: It does not output a result that deteriorates from the original dehazing method result. Also, let us note that we can not hypothesise which is the best original method, as no direct subjective comparison among them was performed. However, we can hypothesise that FVID and DehRet-SRIE are closer to follow the physical model as our method does not present a significant improvement over them.
4 Conclusions
We have presented an approach to induce a physical behaviour to non-physical dehazing methods. Our approach is based on an iterative coupling of the colour channels, which is inspired by the Alternative Least Squares (ALS) method. Results show that our approach is strikingly promising. As further work, we will perform larger experiments with more images and subjects, will consider other evaluation paradigms (e.g. SIFT-based comparison [1]), and will study the convergence of our iterative scheme.
References
Ancuti, C., Ancuti, C.O.: Effective contrast-based dehazing for robust image matching. IEEE Geosci. Remote Sens. Lett. 11(11), 1871–1875 (2014). https://doi.org/10.1109/LGRS.2014.2312314
Ancuti, C., Ancuti, C.: Single image dehazing by multi-scale fusion. IEEE Trans. Image Process. 22(8), 3271–3282 (2013)
Berman, D., Treibitz, T., Avidan, S.: Non-local image dehazing. In: IEEE Conference on Computer Vision and Pattern Recognition (CVPR) (2016)
Cai, B., Xu, X., Jia, K., Qing, C., Tao, D.: DehazeNet: An End-to-End System for Single Image Haze Removal, January 2016. arXiv:1601.07661
Chen, C., Do, M.N., Wang, J.: Robust image and video dehazing with visual artifact suppression via gradient residual minimization. In: Leibe, B., Matas, J., Sebe, N., Welling, M. (eds.) ECCV 2016, Part II. LNCS, vol. 9906, pp. 576–591. Springer, Cham (2016). https://doi.org/10.1007/978-3-319-46475-6_36
Choi, L.K., You, J., Bovik, A.C.: Referenceless prediction of perceptual fog density and perceptual image defogging. IEEE Trans. Image Process. 24(11), 3888–3901 (2015)
Fattal, R.: Dehazing using color-lines. ACM Trans. Graph. 34, 1 (2014)
Finlayson, G.D., Mohammadzadeh Darrodi, M., Mackiewicz, M.: The alternating least squares technique for nonuniform intensity color correction. Color Res. Appl. 40(3), 232–242 (2014). https://doi.org/10.1002/col.21889
Fu, X., Zeng, D., Huang, Y., Zhang, X.P., Ding, X.: A weighted variational model for simultaneous reflectance and illumination estimation. In: 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 2782–2790, June 2016. https://doi.org/10.1109/CVPR.2016.304
Galdran, A., Alvarez-Gila, A., Bria, A., Vazquez-Corral, J., Bertalmío, M.: On the duality between retinex and image dehazing. In: IEEE Conference on Computer Vision and Pattern Recognition (CVPR) (2018)
Galdran, A., Vazquez-Corral, J., Pardo, D., Bertalmío, M.: Enhanced variational image dehazing. SIAM J. Imaging Sci. 8(3), 1519–1546 (2015)
Galdran, A., Vazquez-Corral, J., Pardo, D., Bertalmío, M.: Fusion-based variational image dehazing. IEEE Signal Process. Lett. 24(2), 151–155 (2017). https://doi.org/10.1109/LSP.2016.2643168
He, K., Sun, J., Tang, X.: Single image haze removal using dark channel prior. IEEE Trans. Pattern Anal. Mach. Intell. 33(12), 2341–2353 (2011)
Khoury, J.E., Moan, S.L., Thomas, J., Mansouri, A.: Color and sharpness assessment of single image dehazing. Multimedia Tools Appl. 77(12), 15409–15430 (2018)
Koschmieder, H.: Theorie der horizontalen Sichtweite: Kontrast und Sichtweite. Keim & Nemnich (1925)
Li, Y., Guo, F., Tan, R.T., Brown, M.S.: A contrast enhancement framework with JPEG artifacts suppression. In: Fleet, D., Pajdla, T., Schiele, B., Tuytelaars, T. (eds.) ECCV 2014, Part II. LNCS, vol. 8690, pp. 174–188. Springer, Cham (2014). https://doi.org/10.1007/978-3-319-10605-2_12
Matlin, E., Milanfar, P.: Removal of haze and noise from a single image. In: Proceedings of SPIE 8296. Computational Imaging X, vol. 8296, pp. 82960T–82960T-12 (2012)
Meng, G., Wang, Y., Duan, J., Xiang, S., Pan, C.: Efficient image dehazing with boundary constraint and contextual regularization. In: 2013 IEEE International Conference on Computer Vision (ICCV), pp. 617–624, December 2013
Mittal, A., Moorthy, A.K., Bovik, A.: No-reference image quality assessment in the spatial domain. IEEE Trans. Image Process. 21(12), 4695–4708 (2012). https://doi.org/10.1109/TIP.2012.2214050
Mittal, A., Soundararajan, R., Bovik, A.: Making a “completely blind” image quality analyzer. IEEE Signal Process. Lett. 20, 209–212 (2013)
Nishino, K., Kratz, L., Lombardi, S.: Bayesian defogging. Int. J. Comput. Vis. 98(3), 263–278 (2012)
Petro, A.B., Sbert, C., Morel, J.M.: Multiscale retinex. Image Process. On Line, 71–88 (2014). https://doi.org/10.5201/ipol.2014.107
Tan, R.: Visibility in bad weather from a single image. In: IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2008, pp. 1–8, June 2008
Tarel, J.P., Hautiere, N., Caraffa, L., Cord, A., Halmaoui, H., Gruyer, D.: Vision enhancement in homogeneous and heterogeneous fog. IEEE Intell. Transp. Syst. Mag. 4(2), 6–20 (2012)
Vazquez-Corral, J., Galdran, A., Cyriac, P., Bertalmío, M.: A fast image dehazing method that does not introduce color artifacts. J. Real-Time Image Process., August 2018. https://doi.org/10.1007/s11554-018-0816-6
Wang, S., Cho, W., Jang, J., Abidi, M.A., Paik, J.: Contrast-dependent saturation adjustment for outdoor image enhancement. J. Opt. Soc. Am. A 34(1), 7–17 (2017). https://doi.org/10.1364/JOSAA.34.000007. http://josaa.osa.org/abstract.cfm?URI=josaa-34-1-7
Zhang, H., Patel, V.M.: Densely connected pyramid dehazing network. In: IEEE Conference on Computer Vision and Pattern Recognition (CVPR) (2018)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2019 Springer Nature Switzerland AG
About this paper
Cite this paper
Vazquez-Corral, J., Finlayson, G.D., Bertalmío, M. (2019). Physically Plausible Dehazing for Non-physical Dehazing Algorithms. In: Tominaga, S., Schettini, R., Trémeau, A., Horiuchi, T. (eds) Computational Color Imaging. CCIW 2019. Lecture Notes in Computer Science(), vol 11418. Springer, Cham. https://doi.org/10.1007/978-3-030-13940-7_18
Download citation
DOI: https://doi.org/10.1007/978-3-030-13940-7_18
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-13939-1
Online ISBN: 978-3-030-13940-7
eBook Packages: Computer ScienceComputer Science (R0)