Abstract
The haze model, which describes the degradation of atmospheric visibility, is a good approximation for a wide range of weather conditions and several situations. However, it misrepresents the perceived scenes and causes therefore undesirable results on dehazed images at high densities of fog. In this paper, using data from CHIC database, we investigate the possibility to screen the regions of the hazy image, where the haze model inversion is likely to fail in providing perceptually recognized colors. This study is done upon the perceived correlation between the atmospheric light color and the objects’ colors at various fog densities. Accordingly, at high densities of fog, the colors are badly recovered and do not match the original fog-free image. At low fog densities, the haze model inversion provides acceptable results for a large panel of colors.
You have full access to this open access chapter, Download conference paper PDF
Similar content being viewed by others
1 Introduction
Our atmosphere is filled with particles that cause several phenomena, which lead to visibility degradation and make the detection of scenes’ objects more difficult. The severity of these phenomena is directly related to the type of particles, their size and their concentrations [13].
In such conditions, the apparent luminance of any distant object is controlled by two processes that occur concurrently: the light coming from the object is gradually attenuated by scattering and absorption. It is denoted by t. And the airlight A coming from a source of light (e.g. The Sun) and it is scattered by the haze particles toward the observer all along the line of sight [2].
The formation of the image with the degraded visibility at pixel x, is generally modeled as the sum of the scene’s radiance J(x) and the atmospheric light \(A_{\infty }\), weighted by a transmission factor t(x) (Eq. 1). The \(A_{\infty }\) is the airlight scattered by an object located at infinity with respect to the observer.
where I(x) is the image formed on the camera’s sensor. The transmission factor t(x) depends on the scene’s depth d (distance to the sensor) and on the scattering coefficient \(\beta \) of the haze, such that \(t(x)= e^{-\beta .d(x)}\) [9]. \(t \in [0, 1]\): when \(t=1\), this means that the total light coming from the imaged object reaches the camera with no attenuation. And when \(t=0\), this indicates that the light coming from the imaged object is totally extinct by the fog. In other words, this happens when the fog is opaque.
An important number of dehazing methods consider the inversion of this simple physical model based on prior assumptions to estimate the unknown parameters and to recover J(x) [10, 12]. Some approaches adjust this model to represent faithfully reality, such as adding a parameter to model noise [7]. However, the majority of single image dehazing methods consider the simplest haze model given in 1. These methods consider the haze model evenly all over the image, even when the conditions greatly change between regions, such as the density of fog at distant objects and the color of objects. This seriously degrades the perceptual quality at these regions after dehazing. One may wonder, if this only due to the hypothesis of the used dehazing method or it is also due to the haze model limitation.
For instance, we consider the aerial view shown in Fig. 1: on the lower half part of the image, there is a soft layer of haze, thus, the light transmitted from the objects to the camera is important. The haze-free image recovered through the haze model inversion shows a visually pleasing natural scene. However, on the very upper part where the haze is heavy, the application of the same model inversion provides surprising estimated colors.
Hazy image and the correspondant dehazed image processed by the physical based Dark Channel Prior method [6]. The lower half part of the dehazed image shows a good quality in the objects recovery. However, some colored spots appear on the upper part. Images are borrowed from [10]. (Color figure online)
In order to investigate the haze model validity over various regions of the image, which means the ability to recover recognized colors after inversion, we use data from the CHIC database [5, 8]. It includes, unlike other databases, a set of real foggy scenes with various fog densities in addition to the original fog-free image. It provides also accurate values of the basically unknown parameters of the haze model, t and \(A_{\infty }\). In Sect. 2, we measure the evolution of chromatic dependencies between the airlight and the objects’ colors through the fog thickness. We use the perceptual correlation as a prior screening of regions over which the dehazing is prone to failure. Accordingly, we identify three critical classes for the haze model validity in Sect. 3 based on the perceptual difference between foggy colors and the airlight. This work is concluded in the last section, where we invite researchers to reconsider the haze model at high densities of fog. In the rest of this paper, we use the term dehazing to denote the haze model inversion.
2 Colorimetric Assessment Through Fog Densities
In our previous study [3], using simulated hazy images, the shift of the color components, hue and saturation, was quantified between the haze-free and the dehazed images. At low fog densities, hue is more accurately recovered than saturation. In order to investigate the consistency between real and simulated foggy images, a comparison between their recovered images was conducted in [4]. It was noted that the convergence model fails to stand for high densities of fog and causes inaccuracies in saturation recovery at low densities.
While previous studies focused mainly on the analysis of dehazing results, we investigate here the possibility to screen the inputs where the convergence is prone to failure. The series of these studies serves on the one hand to restrict the dehazing to the proper regions and to warn of the use of simulated images used recently to train neural network models designed for dehazing [1, 11].
Thus, we use two scenes of several densities of fog taken from CHIC database. The images are shown in Fig. 2.
The patches are represented by the corresponding letter in Fig. 4. (Color figure online)
The distribution of r and g coordinates of the ColorChecker patches (Fig. 5) at fog level 7 (red), fog level 9 (green) and fog-free level (blue). The triangle placed at the center is \(A_\infty \). (Color figure online)
2.1 Protocol
We investigate the possibility to locate the foggy colors that might be unrecognized after dehazing according to the perceptual difference between the atmospheric light and the foggy colors. We particularly studied the recovery of the colorchecker patches at fog levels. These colorcheckers are placed at the back of the scenes A and B of the CHIC database at 7 and 4.25 m from the camera, respectively (Fig. 2).
For every color patch indicated by the corresponding letters in Fig. 3, we plot the color coordinates in the rg color space, where r, g and b imply the proportion of red, green and blue in the original RGB color, with \(r + g + b = 1\) (Fig. 4). This visualizes the evolution of colors with the increasing fog density towards the atmospheric light \(A_{\infty }\). Then, for each fog density, we performed the inversion of the haze model to recover the haze-free image J, as the unknown parameters t and \(A_\infty \) are calculable in CHIC database. The recovered images are shown in Fig. 5.
2.2 Indicators
Two parameters are calculated to label the behavior of the haze model across fog densities: the color difference parameter \(\varDelta E^{*}_{ab}\) calculated between the airlight and the foggy colors as an indicator to report a possible inaccurate color recovery. We correlate it to the transmission parameter t, which denotes the intensity of fog according to the relation \(t(x)= e^{-\beta .d(x)}\) defined in Sect. 1. The calculation of this parameter is detailed in [5].
2.3 Analysis
For ease of reading of Fig. 4, we only plot the coordinates of the color patches of the fog-free level, level 9 and level 7. This is quite sufficient to note that the evolution of colors towards \(A_{\infty }\) is not linear. From a certain increasing density of fog, color hue shifts toward \(A_{\infty }\) away from the hue line determined by the coordinates of the same color patch at the fog-free level and the slightly foggy level 9. The shift value and its direction are not consistent across colors. This is mainly dependent on the relative position of the fog-free color to \(A_{\infty }\). Thus, at high fog densities, as there are no constant hue lines, the dehazing in RGB color space, implies a shift in the perceived hue that degrades the quality of the dehazed image.
Besides the distribution of colors in RGB color space in which the dehazing is often performed, we depict the perceptual difference between the original fog-free color and \(A_{\infty }\), denoted by \(\varDelta E^{*}_{ab}\). A higher value means a larger noticeable color difference. This depends simultaneously on the colorfulness of the original color and the intensity of fog.
According to Fig. 5, level 5 of fog densities represents a critical edge: above this level, colors are perceptually well recovered. However, at lower levels, from level 1 to level 4, colors are hardly recognized. This is reflected in the \(\varDelta E^{*}_{ab}\) values given in Tables 1 and 2: considering level 5 for both scenes, A and B, colors are quite better recovered for scene B. The corresponding values of \(\varDelta E^{*}_{ab}\) are higher for scene B than for scene A, which underlines a larger difference against the atmospheric light and therefore a better recovery.
This points out an obvious connection between the initial difference between foggy colors and the atmospheric light, on the one hand, and the quality of dehazed colors on the other hand. According to the values provided in Tables 1 and 2, for values below the color difference \(\varDelta E^{*}_{ab}\) \(\in [5, 6]\), dehazing induces an image quality degaradation. Beyond these values, recovered colors are similar to the fog-free original image and the global quality is improved.
3 Results
In all of this, we can identify three critical classes for the haze model validity according to the perceptual difference parameter \(\varDelta E^{*}_{ab}\) and the transmission parameter t:
-
1.
A high density of fog, which induces small transmission values, makes all colors almost look alike and close to the atmospheric light (from level 1 to level 4 in CHIC database). At these levels, transmission values are extremely low. They vary between 0 and 0.1. The results do not depend on the objects’ colors. No matter what the color is, it is badly recovered by dehazing.
-
2.
A medium density, which shows up when almost \(20\%\) of the transmitted light reach the camera, induces particular observation. Generally, recovered colors are barely recognized, especially those with the lowest \(\varDelta E^{*}_{ab}\) values, such as the color patches B, E and F (see level 5 in Table 1). Moreover, some patches are hardly discernible after dehazing such as patches E and F, and patches K, L and P.
-
3.
A low density of fog, which guarantees the recognition of the color’s hue. From level 6 to level 9, \(\varDelta E^{*}_{ab}\) values either almost fall in the interval [5, 6] or they are higher. Thus, colors are all easily distinguished from the airlight and they are therefore accurately recovered in hue but with a small bias in the colorfulness, which is not quantified in this study. This is mainly due to the noise, which is excluded from the haze model. Transmission values at such levels are higher than 0.2.
4 Conclusion
In this paper, we investigate the possibility to screen regions where the dehazing is likely to fail in providing correct colors that we believe mandatory to avoid adverse effects. Accordingly, beyond a critical density of fog, the perceptual difference between the atmospheric light and the hazy colors is extremely low. Thus, for any color hidden by fog, it is impossible to retrieve the accurate hue of the fog-free color in the visible range, even if the parameters of the fog model are accurately calculated. At lower fog densities, the innacuracy of the color recovery is mainly due to the noise, which affects particularly the colorfulness of the recovered scene.
The physics based dehazing methods used today should consider seriously the haze model validity. Otherwise, undesirable effects show up at very far objects or in the presence of much fog. Usually, we hardly try to reduce these effects by refining hypothesis, while the model itself is limited.
Future works will focus on the quantification of colors’ shift according to the parameters investigated in this paper and the possible coupling of complementary processing operations to push the haze model limits even further.
References
Cai, B., Xiangmin, X., Jia, K., Qing, C., Tao, D.: DehazeNet: an end-to-end system for single image haze removal. IEEE Trans. Image Process. 25(11), 5187–5198 (2016)
Duntley, S.Q.: The reduction of apparent contrast by the atmosphere. JOSA 38(2), 179–191 (1948)
El Khoury, J., Thomas, J.-B., Mansouri, A.: Does dehazing model preserve color information? In: 2014 Tenth International Conference on Signal-Image Technology and Internet-Based Systems (SITIS), pp. 606–613. IEEE (2014)
El Khoury, J., Thomas, J.-B., Mansouri, A.: Haze and convergence models: experimental comparison. In: AIC 2015 (2015)
El Khoury, J., Thomas, J.-B., Mansouri, A.: A color image database for haze model and dehazing methods evaluation. In: Mansouri, A., Nouboud, F., Chalifour, A., Mammass, D., Meunier, J., ElMoataz, A. (eds.) ICISP 2016. LNCS, vol. 9680, pp. 109–117. Springer, Cham (2016). https://doi.org/10.1007/978-3-319-33618-3_12
He, K., Sun, J., Tang, X.: Single image haze removal using dark channel prior. IEEE Trans. Pattern Anal. Mach. Intell. 33(12), 2341–2353 (2011)
Khoury, J.: Model and quality assessment of single image dehazing. Ph.D. thesis, Université de Bourgogne Franche-Comté (2016)
El Khoury, J., Thomas, J.-B., Mansouri, A.: A database with reference for image dehazing evaluation. J. Imaging Sci. Technol. 62, 10503-1–10503-13 (2017)
Koschmieder, H.: Theorie der horizontalen Sichtweite: Kontrast und Sichtweite. Keim & Nemnich, Munich (1925)
Lee, S., Yun, S., Nam, J.-H., Won, C.S., Jung, S.-W.: A review on dark channel prior based image dehazing algorithms. EURASIP J. Image Video Process. 2016(1), 4 (2016)
Li, B., Peng, X., Wang, Z., Xu, J., Feng, D.: AOD-Net: all-in-one dehazing network. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 4770–4778 (2017)
Li, Y., You, S., Brown, M.S., Tan, R.T.: Haze visibility enhancement: a survey and quantitative benchmarking. Comput. Vis. Image Underst. 165, 1–16 (2017)
McCartney, E.J.: Optics of the Atmosphere Scattering by Molecules and Particles. Wiley, New York (1976). 421 p
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2018 Springer International Publishing AG, part of Springer Nature
About this paper
Cite this paper
El Khoury, J., Thomas, JB., Mansouri, A. (2018). Colorimetric Screening of the Haze Model Limits. In: Mansouri, A., El Moataz, A., Nouboud, F., Mammass, D. (eds) Image and Signal Processing. ICISP 2018. Lecture Notes in Computer Science(), vol 10884. Springer, Cham. https://doi.org/10.1007/978-3-319-94211-7_52
Download citation
DOI: https://doi.org/10.1007/978-3-319-94211-7_52
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-94210-0
Online ISBN: 978-3-319-94211-7
eBook Packages: Computer ScienceComputer Science (R0)