Abstract
Boosting is a classic algorithm which has been successfully applied to diverse computer vision tasks. In the scenario of image denoising, however, the existing boosting algorithms are surpassed by the emerging learning-based models. In this paper, we propose a novel deep boosting framework (DBF) for denoising, which integrates several convolutional networks in a feed-forward fashion. Along with the integrated networks, however, the depth of the boosting framework is substantially increased, which brings difficulty to training. To solve this problem, we introduce the concept of dense connection that overcomes the vanishing of gradients during training. Furthermore, we propose a path-widening fusion scheme cooperated with the dilated convolution to derive a lightweight yet efficient convolutional network as the boosting unit, named Dilated Dense Fusion Network (DDFN). Comprehensive experiments demonstrate that our DBF outperforms existing methods on widely used benchmarks, in terms of different denoising tasks.
You have full access to this open access chapter, Download conference paper PDF
Similar content being viewed by others
Keywords
These keywords were added by machine and not by the authors. This process is experimental and the keywords may be updated as the learning algorithm improves.
1 Introduction
Image denoising is a typical problem in low-level computer vision. Observed a contaminated image with a certain kind of noise (e.g., additive white Gaussian noise), plenty of methods have been investigated to restore the original signal. Among them, modeling image priors for restoration is a prominent approach, such as nonlocal similarity based models [1,2,3] and sparsity based models [4,5,6]. Specifically, BM3D [7], CSF [8], and WNNM [9] are several representative methods for image denoising.
Recently, with the rapid advancement of GPU-based parallel computing frameworks, increasingly more learning-based denoising models [10,11,12,13] began to adopt the paradigm of end-to-end training based on a convolutional neural network (CNN). These learning-based models have achieved competitive or even better performance than previous methods. On the other hand, several traditional models [14,15,16] based on the boosting algorithm studied the denoising problem in a unique perspective. By extracting the residual signal or eliminating the noise leftover, these methods boost the restoration quality iteratively. Beyond them, Romano and Elad proposed a notable variant of the boosting algorithm, named Strengthen-Operate-Subtract (SOS) [17]. By combining the denoised image with the original input, it increased the signal-to-noise ratio iteratively and achieved promising improvements.
Nevertheless, the existing boosting algorithms still have performance gaps compared with the learning-based models. In this paper, we embed the deep learning technique into the boosting algorithm and significantly boost its performance in the scenario of image denoising. Specifically, we construct a Deep Boosting Framework (DBF) that integrates several CNNs in a feed-forward fashion, where each network serves as a boosting unit. To the best of our knowledge, this is the first time that deep learning and boosting are jointly investigated for image restoration. Although this paper mainly focuses on denoising, the proposed DBF can be readily generalized to other image restoration tasks.
Theoretically, the boosting unit in the DBF can be any type of networks. In practice, however, we find that not all structures are suitable to be employed as the boosting unit. The reason is that, along with the integrated networks for boosting, the depth of the DBF is substantially increased. It thus causes the difficulty of convergence during training. To fully exploit the potential of the DBF, we further propose a Dilated Dense Fusion Network (DDFN), which is highly optimized to serve as the boosting unit.
We reform the plainly connected network structure in three steps to obtain the DDFN. First, to overcome the vanishing of gradients during training, we introduce the dense connection [18] to construct the boosting unit, which also improves the re-usage of features and thus guarantees the efficiency. Second, to obtain better performance based on the densely connected structure, we adopt the dilated convolution [19] for widening the receptive field without additive parameters, which maintains the lightweight structure of the boosting unit. Last but not least, we further propose a path-widening fusion scheme cooperated with the dilated convolution to make the boosting unit more efficient.
The contributions of this paper are summarized as follows:
(1) We propose a novel boosting framework for image denoising by introducing deep learning into the boosting algorithm, named DBF. It not only outperforms existing boosting algorithms by a large margin but also performs better than extensive learning-based models.
(2) We optimize a lightweight yet efficient convolutional network as the boosting unit, named DDFN. With the dense connection, we address the difficulty of convergence in DBF. Cooperating with the dilated convolution, we propose a path-widening fusion scheme to expand the capacity of each boosting unit.
(3) Our DDFN-based DBF has a clear advantage over existing methods on widely used benchmarks when trained at a specific noise level. If trained for blind Gaussian denoising, it achieves a new state-of-the-art performance within a wide range of noise levels. Also, the proposed method is demonstrated effective when generalized to the image deblocking task.
2 Related Work
CNN-Based Image Denoising. Research along this direction focuses on the exploration of the network structure. Advanced design of architecture yields better restoration quality. For instance, Burger et al. [10] trained a multi-layer perceptrons (MLPs) with a large image database, which achieved comparable results with BM3D [7]. Chen et al. [11] proposed a stage-wise model (i.e., TNRD) which introduced the well-designed convolutional layers into the non-linear diffusion model to derive a flexible framework. And Zhang et al. [12] composed the deep DnCNN model by utilizing batch normalization (BN) [20] and residual connection [21]. Essentially, DnCNN can be viewed as the generalization of one-stage TNRD. Besides, a recently proposed model [13] combined image denoising with semantic classification using CNNs, which bridged the gap between these two different tasks and improved the restoration quality. Following the successful paradigm of end-to-end training, we also adopt the CNN for image denoising. Different from its common usages, however, the employed CNN is just a component of our denoising model. Specifically, it is integrated as a boosting unit in the DBF. Experimental results demonstrate the superior performance of our boosting framework compared with a single CNN model.
Boosting Algorithm. Boosting is a widely used algorithm to improve the performance of diverse tasks by cascading several steerable sub-models. A plenty of models based on the boosting algorithm have been investigated for image denoising in literature [14,15,16, 22]. Generally, the detailed implementation can be divided into 3 classes: (a) re-utilizing the residual [14], (b) re-enhancing the denoised signal [15], and (c) strengthening the SNR iteratively [17]. However, these boosting algorithms with classic models are surpassed by the emerging learning-based models. Contrastively, our proposed DBF inherits both advantages of boosting and CNN and achieves a new state-of-the-art performance for image denoising. Note that, boosting and CNN have been combined for image classification tasks before, e.g., IB-CNN [23] and BoostCNN [24], yet our proposed DBF is the first deep boosting framework in the field of image restoration.
3 Deep Boosting Framework
3.1 Boosting Perspective of Denoising
The fundamental image denoising problem is the recovery of an image \(x\in \mathbb {R}^{N\times M}\) from a contaminated measurement \(y\in \mathbb {R}^{N\times M}\), which can be formulated as
where v stands for the additive noise that is generally modeled as zero-mean white Gaussian noise with a standard deviation \(\sigma \). The denoising process can be represented as
where the operator \(S(\cdot )\) stands for a general denoising method and \(\hat{x}\) stands for an approximation of x.
However, the image \(\hat{x}\) recovered by any algorithm cannot ideally equal to x, and the gap between them can be denoted as
where \(x_r\) represents the unrecovered signal and \(v_r\) stands for the leftover noise in \(\hat{x}\). In other words, by adding \(x_r\) and subtracting \(v_r\), we then obtain the clean image x from \(\hat{x}\).
A straightforward idea to apply the boosting algorithm to image denoising is that we iteratively extract the unrecovered signal \(x_r\) from the residual and add it back to \(\hat{x}\)
where \(\mathcal {H}(\cdot )\) is an operator for the iterative extraction and we set \(\hat{x}^{0}=0\). Note that, however, the residual \(y-\hat{x}\) contains not only the unrecovered signal \(x_r\) but also a part of noise
Another native idea is that we remove the leftover noise \(v_r\) by filtering the denoised image \(\hat{x}\) iteratively
where \(\mathcal {F}(\cdot )\) stands for a certain denoising model. However, it could lead to over-smoothing since it neglects \(x_r\) which contains most high frequency information.
To further improve the performance of the boosting framework, Romano and Elad proposed a novel SOS algorithm. The denoising target in each iteration step is the “strengthened” image \(y+\hat{x}^n\), instead of the residual \(y-\hat{x}^n\) or the denoised image \(\hat{x}^n\), which improves the signal-to-noise ratio (SNR) [17]. To guarantee the iterability of SOS, however, it has to “subtract” the identical \(\hat{x}^n\) in each step as
where \(\mathcal {G}(\cdot )\) is a certain denoising model imposed on the strengthened image. To better clarify the insight of the SOS algorithm, we decompose \(y+\hat{x}\) as
Assuming that \(||u||=\delta ||v||\), where \(\delta \ll 1\). Then we have \(SNR(y+\hat{x})>SNR(y)\) according to the Cauchy-Schwarz inequality [17]. All we need to achieve so is a general denoising method even if it is a “weak” one.
3.2 CNN-Based Boosting
Inspired by SOS [17], we propose a new boosting framework by leveraging deep learning. Specifically, we introduce a CNN to learn the denoising model in each stage. Following Eq. (7), we have
where \(\theta \) denotes the trainable parameter set of the CNN.
The subtraction of identical \(\hat{x}^{n}\) inherited from Eq. (7) aims to guarantee the iterability of the SOS algorithm. Such constraint in Eq. (9) is no longer needed since we can learn different denoising models in each stage. In other words, our deep boosting framework can adjust its parameters without the constraint of identical subtraction, which actually yields a better performance as will be demonstrated in Sect. 5.3. The output of the final stage can be represented as
where n stands for the serial number of each stage. Figure 1 illustrates a flowchart for Eq. (10) for a better understanding.
The loss function for training the parameters \(\varTheta =\{\theta _1,\theta _2,...,\theta _n\}\) is the mean square error (MSE) between the final output \(\hat{x}^{n}\) and the ground truth x
where B denotes the size of mini-batch for the stochastic gradient descent. Such training scheme is called joint training which optimizes the parameters in all stages simultaneously. We also consider a greedy training scheme, for which the parameters are firstly optimized stage-wise and then fine-tuned among all stages. Related experimental results will be described in Sect. 5.3.
CNN-based deep boosting framework. The B.Unit\(_n\) denotes the \(n^{th}\) boosting unit (i.e., \(\mathcal {G}_{\theta _n}\)) in the framework. The investigation of B.Unit is detailed in Sect. 4
3.3 Relationship to TNRD
The TNRD model proposed in [11] is also a stage-wise model trained jointly, which can be formulated as
where \(\mathcal {D}(\cdot )\) stands for the diffusion term which is implemented using a CNN with two layers and \(\mathcal {R}(\cdot )\) denotes the reaction term as \(\mathcal {R}(\hat{x}^{n-1}, y)=\gamma (\hat{x}^{n-1}-y)\), where \(\gamma \) is a factor which denotes the strength of the reaction term.
Actually, TNRD can be interpreted as a special case of the boosting algorithm. Combining Eqs. (4) and (6), we have
Providing \(\mathcal {F}(\cdot )=-\mathcal {D}(\cdot )\) and \(\mathcal {H}(\cdot )=-\gamma (\cdot )\), we then obtain the basic equation of the TNRD model.
However, by further decomposing Eq. (12), we demonstrate the fundamentally different insights between TNRD and DBF as follows. Without loss of generality, let \(\hat{x}^{n-1}=x+u\) and we discuss a special case when \(\gamma =1\). Considering Eqs. (1) and (3), we have
Substituting Eq. (14) into Eq. (12), we then have
Details of the evolution for the boosting unit. “C” and “D” with a rectangular block denote the convolution and its dilated variant, respectively. The following “1” and “3” denote the kernel size. “+” with a circular block denotes the concatenation. Each layer in DDFN (except the last one) adopts ReLU [25] as the activation function, which is omitted here for simplifying the illustration
The target of TNRD is to let \(\hat{x}^{n}\rightarrow x\), i.e., \(\mathcal {D}(\hat{x}^{n-1})\rightarrow v\). Thus, the diffusion term is actually trained for fitting the white Gaussian noise v. In contrast, our proposed DBF is trained for directly restoring the original signal x, leveraging on the availability of denoised images and the growth of SNR. Intuitively, it may be more difficult to find correlations between training examples when fitting the irregular noise. Moreover, from the perspective of SNR, it is more difficult to predict the “weaker” noise when the input image has a lower noise level. These are the advantages of our DBF in comparison with TNRD.
4 Dilated Dense Fusion Network
An efficient boosting unit is desired to fully exploit the potential of the proposed DBF. Theoretically, the function \(\mathcal {G}_\theta (\cdot )\) in Eq. (10) has no restriction on the detailed implementation of the boosting unit. Thus, we have a wide choice of diverse network structures. We start our investigation from a simple structure which is the simplified DnCNN [12] without batch normalization and residual connection, i.e., the plain network (PN), as shown in Table 1 and Fig. 2(a). We find in experiments that, given the same number of parameters, deepening a network properly contributes to the efficiency (as detailed in Sect. 5.2). However, when we introduce the PN into our DBF to derive a 2-stage boosting framework, this benefit tends to vanish as the network depth continues to increase, probably due to the vanishing of gradient during the back propagation.
4.1 Dense Connection
To overcome the propagation problem of gradient during training, we introduce the dense connection to derive the dense network (DN), as shown in Table 1 and Fig. 2(b), which is inspired by the successful model for image recognition [18]. The dense connection enables the \(l^{th}\) layer to receive the features of all preceding layers (i.e., \(f_0,...,f_{l-1}\)) as input
where \(g_l(\cdot )\) denotes the \(l^{th}\) layer in \(\mathcal {G}_\theta \) and \([f_0,f_1,...,f_{l-1}]\) stands for the concatenation of the features output from preceding layers. We demonstrate in experiments that the dense connection can address the propagation issue of gradient during training (as detailed in Sect. 5.2).
4.2 Dilated Convolution
Widening the receptive field of the CNN is a well-known strategy for enhancing the performance in both image classification [26] and restoration [27] tasks. The convolution with a larger kernel size can widen the receptive field, however, it increases the number of parameters at the same time. Another strategy is stacking multiple convolutional layers with a \(3\times 3\) kernel size to obtain a large receptive field equivalently. However, it causes difficulty of convergence due to the increasing of the network depth.
Recently, a notable alternative called dilated convolution has been investigated in semantic segmentation [19] and image classification [28]. The dilated convolution can widen the receptive field without additive parameters and it also prevents the increasing of depth. Inspired by that, we introduce the dilated convolution to derive the dilated dense network (DDN) based on the DN, as shown in Table 1 and Fig. 2(c). By widening the receptive field efficiently, a better denoising performance can be achieved (as detailed in Sect. 5.2).
4.3 Path-Widening Fusion
We further propose a path-widening fusion scheme to make the boosting unit more efficient. As shown in Table 1 and Fig. 2(d), we expand the number of forward paths to derive the DDFN from the DDN. Specifically, in a certain block, the order between the dilated convolutions (Dconv for short) and the normal convolutions (Conv for short) is exchanged in different branches. It is very likely that the Conv-ReLU-Dconv and Dconv-ReLU-Conv branches can learn different feature representations. The proposed path-widening fusion exploits the potential of these two orders at the same time, and thus promotes the possibility to learn better representations. Experimental results demonstrate that the denoising performance can be further improved in this way (as detailed in Sect. 5.2). Note that, we restrict the parameter number of DDFN not greater than DDN (i.e., about \(4\times 10^4\)) to eliminate the influence of additional parameters due to path-widening fusion, and thus the efficiency of DDFN is also justified.
5 Experimental Results
5.1 Datasets and Settings
We adopt 400 images at a \(180\times 180\) resolution for training our models following TNRD [11] and DnCNN [12]. The images are partitioned into sub-image patches with a size of \(50\times 50\), and the mini-batch number is set to 64 for the stochastic gradient decent. Two widely-used datasets, “Set12” and “BSD68” [29] are employed as the benchmarks for image denoising. Moreover, to compare with the SOS algorithm [17], the “Set5” dataset is adopted following [17].
Besides grey-level image denoising, we also apply our method to two additional tasks, i.e., color image denoising and JPEG image deblocking, following the setting of DnCNN [12]. The color version of “BSD68” is adopted for the color image denoising task. And the “Classic5” and “LIVE1” datasets are adopted for evaluating the deblocking task as in [30].
We use TensorFlow and the “Adam” [31] solver for optimization with the momentum factor set as 0.9 and the coefficient of weight decay (\(L_2\)-Norm) as 0.0001. The learning rate is decayed exponentially from 0.001 to 0.0001. We stop training when no notable decay of training loss is observed after \(3.6\times 10^5\) iterations. The algorithm proposed in [32] is adopted for initializing the weights except the last layer. Specifically, the last layer for reconstruction is initialized by the random weights drawn from Gaussian distributions with \(\sigma =0.001\). And we set zeros for initializing the biases in each convolutional layer.
Illustrations of the ablation experiments. (a) The curves show the advantage of dense connection over plain structure in terms of convergence. (b) The evolution from plainly connected structure to DDFN. (c) Performance comparisons between DBF and its variants. The symbol “W” means wide and all of these models are tested on the “Set12” dataset at \(\sigma =50\)
5.2 Ablation Experiments of DDFN
The proposed DDFN integrates three concepts: dense connection, dilated convolution, and path-widening fusion, deriving a fundamentally different structure compared with existing models. In this section, we design extensive ablation experiments to evaluate them respectively.
Investigation of Depth (PN). We described the structure of PN in Sect. 4. To investigate the effect of depth to the boosting unit, we construct a variant of PN (named PN2) with a deeper yet thinner structure, which has the same number of parameters compared with PN. Specifically, it contains more layers (i.e., 18) in the feature integration part and less filter numbers (i.e., 16) in each layer than PN. Meanwhile, we keep the other hyper-parameters and the training procedure of PN2 the same as PN. As shown in Fig. 3(b), the deeper and thinner PN2 outperforms PN. This observation suggests that deepening the framework gives a better performance. However, when we introduce PN2 into DBF to derive PN2-x2, the advantage of plainly connected deeper structure tends to vanish.
Dense Connection (DN). We then introduce the dense connection to address the propagation issue of gradient during training. As shown in Fig. 3(a), DN converges faster than PN2. While maintaining a quicker convergence, the derived DN-x2 shows a clear advantage over PN2-x2 for a 2-stage DBF, as shown in Fig. 3(b). Note that, the parameters of DN are 15% less than PN2, yet DN-x2 still outperforms PN2-x2.
Dilated Convolution (DDN). Based on DN, we adopt the dilated convolution to widen the receptive field. Specifically, we introduce it into two places of the network: the feature extraction part and each dense block, as shown in Fig. 2(c). The ratio of dilation is fixed to 2 for each dilated convolution layer. Experimental results demonstrate that further improvements of the boosting unit can be achieved, i.e., DDN as shown in Fig. 3(b).
Path-Widening Fusion (DDFN). As described in Sect. 4, we further propose the path-widening fusion which aggregates the concatenated features of preceding layers using a \(1\times 1\) convolution in the dense block, as shown in Fig. 2(d). This fusion can further promote the denoising performance, i.e., DDFN as shown in Fig. 3(b).
5.3 Investigation of Framework
Ablation of Subtraction and Training Scheme. As described in Sect. 3.2, the proposed DBF no longer needs an subtraction of \(\hat{x}^n\) as in SOS to guarantee the iterability. We design an ablation experiment based on a 3-stage DBF. Experimental results demonstrate a better performance (+0.12 dB) without the subtraction. As for the training scheme, we consider both joint and greedy training as described in Sect. 3.2. Evaluated on a 3-stage DBF, we find joint training and greedy training give competitive performance.
Boosting - The Deeper, the Better. We investigate the performance by increasing the stage number of DBF. Experimental results demonstrate the capacity of our DBF in term of the extension in depth, as can be observed from Fig. 3(c). Specifically, a 5-stage DBF brings 0.30 dB gain compared with a single stage one.
DDFN - The Wider, the Better. Besides the exploration of depth, we also investigate the contribution of width by doubling the number of filters in each layer of DDFN (deriving models with the symbol “W”). Experimental results in Fig. 3(c) demonstrate that widening can further enhance the performance.
5.4 Comparison with State-of-the-Art Methods
Comparison with the SOS Algorithm. We adopt four classical models [1, 2, 5, 7] and their corresponding SOS [17] variants for comparison. As shown in Table 2, our boosting unit DDFN has a clear advantage over these classic models, e.g., +0.46 dB when \(\sigma =25\) and +0.42 dB when \(\sigma =50\) than BM3D [7]) on the “Set5” dataset. With the proposed DBF, our DDFN-x5W achieves notable improvements over BM3D-SOS [17], i.e., +0.77 dB (\(\sigma \) = 25) and +0.92 dB (\(\sigma \) = 50).
Comparison with Other Learning-Based Models. We adopt seven representative models for comparison: MLP [10], CSF [8], GCRF [33], TNRD [11], NLNet [34], DeepAM [35] and DnCNN [12]. The restoration results of our DDFN and DDFN-x5W are listed in Table 3 to compare with them. Specifically, DDFN-x5W achieves a superior performance than TNRD [11] (+0.65 dB) and DnCNN [12] (+0.28 dB) on the “Set12” dataset when \(\sigma =50\).
Comparison for Blind Gaussian Denoising. Following the settings of training proposed in [12], we re-train our models to derive the DDFN-B and DDFN-x5W-B for blind Gaussian denoising. We adopt the BM3D [7] and the variant of DnCNN (i.e., DnCNN-B [12]) for comparison. Experimental results listed in Table 4 demonstrate the superiority of our model within a wide range of noise levels. Specifically, when the noise level is small (i.e., \(\sigma =5\)), our proposed DDFN-x5W-B has a clear advantage (+0.44 dB) over DnCNN-B. We also evaluate the performance on the task of blind color image denoising. Experimental results listed in Table 5 demonstrate the advantage of our proposed DBF.
Comparison for JPEG Image Deblocking. We also evaluate our model on the task of image deblocking. Three representative models: AR-CNN [30], TNRD [11], and DnCNN [12] are adopted for comparison. Experimental results listed in Table 6 demonstrate the superiority of our model over existing ones.
Running Time. Although the cascaded structure of our proposed DBF involves more computation than a single stage one (which is the inevitable cost of boosting), it is still quite efficient. Detailed results are listed in Table 7.
Visual Comparison. To evaluate the perceptual quality of restoration, we show a few examples including grey-level denoising (Fig. 4), blind color image denoising (Figs. 5 and 6), and image deblocking (Figs. 7 and 8). As can be seen, our model performs better than the competitors in both the smooth and edge regions.
6 Conclusions
In this paper, we propose the DBF which first integrates the boosting algorithm with deep learning for image denoising. To fully exploit the potential of this framework, we elaborate the lightweight yet efficient DDFN as the boosting unit. By introducing the dense connection, we address the vanishing of gradients during training. Based on the densely connected structure, we further propose the path-widening fusion cooperated with the dilated convolution to optimize the DDFN for efficiency. Compared with the existing models, our DDFN-based DBF achieves the state-of-the-art performance in both non-blind and blind image denoising on widely used benchmarks.
Besides the scenario of image denoising, the proposed DBF can be readily generalized to other image restoration tasks, e.g., image deblocking, as demonstrated in this paper. Also, the idea of path-widening fusion is demonstrated to be useful in the task of spectral reconstruction from RGB images [37]. We believe the proposed method could inspire even more low-level vision applications.
References
Zoran, D., Weiss, Y.: From learning models of natural image patches to whole image restoration. In: IEEE International Conference on Computer Vision (ICCV), pp. 479–486 (2011)
Buades, A., Coll, B., Morel, J.M.: A non-local algorithm for image denoising. In: IEEE Conference on Computer Vision and Pattern Recognition (CVPR), vol. 2, pp. 60–65, June 2005
Buades, A., Coll, B., Morel, J.: Nonlocal image and movie denoising. Int. J. Comput. Vis. (IJCV) 76(2), 123–139 (2008)
Mairal, J., Bach, F., Ponce, J., Sapiro, G., Zisserman, A.: Non-local sparse models for image restoration. In: IEEE International Conference on Computer Vision (ICCV), pp. 2272–2279 (2009)
Elad, M., Aharon, M.: Image denoising via sparse and redundant representations over learned dictionaries. IEEE Trans. Image Process. 15(12), 3736–3745 (2006)
Dong, W., Zhang, L., Shi, G., Li, X.: Nonlocally centralized sparse representation for image restoration. IEEE Trans. Image Process. 22(4), 1620–1630 (2013)
Dabov, K., Foi, A., Katkovnik, V., Egiazarian, K.: Image denoising by sparse 3-D transform-domain collaborative filtering. IEEE Trans. Image Process. 16(8), 2080–2095 (2007)
Schmidt, U., Roth, S.: Shrinkage fields for effective image restoration. In: IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 2774–2781, June 2014
Gu, S., Zhang, L., Zuo, W., Feng, X.: Weighted nuclear norm minimization with application to image denoising. In: IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 2862–2869, June 2014
Burger, H.C., Schuler, C.J., Harmeling, S.: Image denoising: can plain neural networks compete with BM3D? In: IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 2392–2399, June 2012
Chen, Y., Pock, T.: Trainable nonlinear reaction diffusion: a flexible framework for fast and effective image restoration. IEEE Trans. Patt. Anal. Mach. Intell. 39(6), 1256 (2017)
Zhang, K., Zuo, W., Chen, Y., Meng, D., Zhang, L.: Beyond a gaussian denoiser: residual learning of deep CNN for image denoising. IEEE Trans. Image Process. 26(7), 3142–3155 (2017)
Remez, T., Litany, O., Giryes, R., Bronstein, A.M.: Deep class-aware image denoising. In: International Conference on Sampling Theory and Applications (SampTA), pp. 138–142, July 2017
Charest, M.R., Elad, M., Milanfar, P.: A general iterative regularization framework for image denoising. In: Conference on Information Sciences and Systems, pp. 452–457, March 2006
Milanfar, P.: A tour of modern image filtering: new insights and methods, both practical and theoretical. IEEE Sig. Process. Mag. 30(1), 106–128 (2013)
Talebi, H., Zhu, X., Milanfar, P.: How to SAIF-ly boost denoising performance. IEEE Trans. Image Process. 22(4), 1470–1485 (2013)
Romano, Y., Elad, M.: Boosting of image denoising algorithms. SIAM J. Imaging Sci. 8(2), 1187–1219 (2015)
Huang, G., Liu, Z., van der Maaten, L., Weinberger, K.Q.: Densely connected convolutional networks. In: IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 2261–2269, July 2017
Yu, F., Koltun, V.: Multi-scale context aggregation by dilated convolutions. In: International Conference on Learning Representations (ICLR) (2016)
Ioffe, S., Szegedy, C.: Batch normalization: accelerating deep network training by reducing internal covariate shift. In: International Conference on Machine Learning (ICML), pp. 448–456 (2015)
He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 770–778, June 2016
Buades, A., Coll, B., Morel, J.M.: A review of image denoising algorithms, with a new one. SIAM J. Multiscale Model. Simul. 4(2), 490–530 (2005)
Han, S., Meng, Z., Khan, A.S., Tong, Y.: Incremental boosting convolutional neural network for facial action unit recognition. In: Lee, D.D., Sugiyama, M., Luxburg, U.V., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems (NIPS), pp. 109–117. Curran Associates, Inc. (2016)
Moghimi, M., Saberian, M., Yang, J., Li, L.J., Vasconcelos, N., Belongie, S.: Boosted convolutional neural networks. In: British Machine Vision Conference (BMVC), pp. 24.1–24.13 (2016)
Nair, V., Hinton, G.E.: Rectified linear units improve restricted Boltzmann machines. In: International Conference on Machine Learning (ICML), pp. 807–814 (2010)
Szegedy, C., Vanhoucke, V., Ioffe, S., Shlens, J., Wojna, Z.: Rethinking the inception architecture for computer vision. In: IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 2818–2826, June 2016
Kim, J., Lee, J.K., Lee, K.M.: Accurate image super-resolution using very deep convolutional networks. In: IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 1646–1654, June 2016
Yu, F., Koltun, V., Funkhouser, T.: Dilated residual networks. In: IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 636–644, July 2017
Roth, S., Black, M.J.: Fields of experts: a framework for learning image priors. In: IEEE Conference on Computer Vision and Pattern Recognition (CVPR), vol. 2, pp. 860–867, June 2005
Dong, C., Deng, Y., Loy, C.C., Tang, X.: Compression artifacts reduction by a deep convolutional network. In: IEEE International Conference on Computer Vision (ICCV), pp. 576–584, December 2015
Kingma, D., Ba, J.: Adam: a method for stochastic optimization. In: International Conference on Learning Representations (ICLR), vol. 5 (2015)
He, K., Zhang, X., Ren, S., Sun, J.: Delving deep into rectifiers: surpassing human-level performance on ImageNet classification. In: IEEE International Conference on Computer Vision (ICCV), pp. 1026–1034, December 2015
Vemulapalli, R., Tuzel, O., Liu, M.Y.: Deep Gaussian conditional random field network: a model-based deep network for discriminative denoising. In: IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 4801–4809, June 2016
Lefkimmiatis, S.: Non-local color image denoising with convolutional neural networks. In: IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 5882–5891, July 2017
Kim, Y., Jung, H., Min, D., Sohn, K.: Deeply aggregated alternating minimization for image restoration. In: IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 284–292, July 2017
Dabov, K., Foi, A., Katkovnik, V., Egiazarian, K.: Color image denoising via sparse 3D collaborative filtering with grouping constraint in luminance-chrominance space. In: IEEE International Conference on Image Processing (ICIP), vol. 1, I-313–I-316, September 2007
Shi, Z., Chen, C., Xiong, Z., Liu, D., Wu, F.: HSCNN+: advanced CNN-based hyperspectral recovery from RGB images. In: IEEE Conference on Computer Vision and Pattern Recognition (CVPR) Workshops, June 2018
Acknowledgements
We acknowledge funding from National Key R&D Program of China under Grant 2017YFA0700800, and Natural Science Foundation of China under Grants 61671419 and 61425026.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2018 Springer Nature Switzerland AG
About this paper
Cite this paper
Chen, C., Xiong, Z., Tian, X., Wu, F. (2018). Deep Boosting for Image Denoising. In: Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y. (eds) Computer Vision – ECCV 2018. ECCV 2018. Lecture Notes in Computer Science(), vol 11215. Springer, Cham. https://doi.org/10.1007/978-3-030-01252-6_1
Download citation
DOI: https://doi.org/10.1007/978-3-030-01252-6_1
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-01251-9
Online ISBN: 978-3-030-01252-6
eBook Packages: Computer ScienceComputer Science (R0)