1 Introduction
Reproducing a high quality image faithful to the scene is an essential goal of digital photography. The real images, however, are often blurred during image acquisition due to the effect of many factors such as camera shake, object motion, and outoffocus [27]. The resulting blurry images will not only degrade the perceptual quality of photos but also degenerate the performance of many image analytic and understanding models [22]. Blind image deblurring, which has been studied extensively in low level vision for decades of years [25], plays an essential role in improving the visual quality of realworld blurry images.
In general, the purpose of blind image deblurring is to recover the latent sharp image y from its blurry observation x = k y + n, where k is an unknown blur kernel (i.e., uniform or nonuniform), n is an additive white Gaussian noise and denotes the convolution operator. This inverse problem, however, is severely illposed and requires extra information on latent image y to constrain the solution space. Thus, there are two categories of approaches for utilizing prior knowledge, i.e
., optimizationbased and deep learning based deblurring methods. Optimizationbased approaches explicitly model prior knowledge to regularize the solution space of blur kernel
[4, 20, 28, 37, 46] and latent image [6, 2, 30, 42, 24]. In contrast, deep learning based methods [27, 22, 38, 43] implicitly utilize prior knowledge by learning a direct mapping (e.g., convolutional neural network, CNN) from degraded image to latent clean image.For blind image deblurring problem, optimizationbased and deep learning methods respectively have their merits and limitations. Optimizationbased methods are flexible in incorporating versatile prior or regularization [4, 20, 46, 30, 42] tailored for blind deblurring, but suffer from the timeconsuming optimization procedure and oversimplified assumptions on blur kernel (e.g., spatially invariant and uniform). Moreover, conventional image priors (e.g., total variation [4]) are limited in blind deblurring and prone to ordinary solution of delta kernel. Specific priors, e.g., norm [41] and normalized sparsity [20]
, are then suggested for blur kernel estimation. On the other hand, deep learning methods
[27, 22, 38, 43] benefiting from endtoend training and joint optimization can enjoy a fast speed and flexibility in handling spatially variant blur in the dynamic scene. However, deep models learn the direct mapping for blind deblurring, and may be limited in capturing specific priors for blind deblurring. As for dynamic scene deblurring, existing dataset [27] is of relatively small scale, which may be a factor hindering the performance of learned model.Taking the merits and drawbacks of optimizationbased and deep learning based methods into account, one interesting question is to ask whether we can exploit prior model to constrain both the network architecture and learning losses for improved dynamic scene deblurring performance. In this paper, we make the first attempt to address this challenging problem. In particular, based on the effectiveness of image prior in blind deblurring, we propose an Extreme Channel Prior embedded Network (ECPeNet) to help the restoration of latent clean image. The critical component of ECPeNet is a novel trainable extreme channel prior embedded layer (ECPeL), which can aggregate extreme channel and blurry image representations to leverage their respective advantages. By enforcing sparsity on both dark and bright channel of feature maps, we can regularize the solution space of CNN during training, thereby incorporating extreme channel priors into ECPeNet.
Furthermore, existing deep dynamic scene deblurring models [27, 38] usually adopt multiscale network architecture but only consider the coarsetofine information flow. That is, blind deblurring is first performed at the small scale, and then deblurring results (or latent representations) are combined with feature representations of a larger scale for further refinement. However, we show that feature representations of a larger scale actually also benefit the dynamic scene deblurring at a smaller scale. To this end, we present a more effective multiscale network architecture that works in both coarsetofine and finetocoarse manners for better exploiting information flow across scales. Experimental results on GoPro and Köhler datasets show that our proposed ECPeNet performs favorably against stateoftheart deep models. With the above claimed advantages, the ECPeNet can produce visually pleasing results, as shown in Figure 1.
The contribution of this paper is threefold: (i) We propose a novel trainable structural layer, which can aggregate both data information and prior knowledge together to leverage their respective merits but avoid limitations. To the best of our knowledge, this is the first attempt to plug prior knowledge (i.e., statistical properties) into a deblurring network in an endtoend manner. (ii) We introduce a new multiscale network architecture to fully exploit different resolution images for maximizing the information flow. (iii) Extensive experimental results on GoPro [27] and Köhler [18] datasets demonstrate that our ECPeNet outperforms stateoftheart dynamic scene deblurring methods.
2 Related Work
In this section, we briefly review the recent optimizationbased and deep learning based image deblurring methods.
2.1 Optimizationbased Deblurring Methods
The optimizationbased methods aim to develop effective image priors to favor clean images over the blurry one. Representative priors include sparse gradients [6, 33, 23, 40], hyperLaplacian prior [19], normalized sparsity prior [20], Lnorm prior [41], patch recurrence prior [26] and discriminative learned prior [46, 24]. By taking advantages of the aforementioned priors, existing optimizationbased methods could deliver competitive results on generic natural images. These approaches, however, cannot be generalized well to handle domain specific images. Thus, specific priors are needed to be introduced for specific images (e.g., a light streak prior [13] for low light images, and a combination of intensity and gradient prior [29] for text images). Recently, Pan et al. [30] developed a dark channel prior (DCP) [10] based model to enforce sparsity on the dark channel of latent image and achieved promising result on both generic and specific images. With the success of [30], Yan et al. [42] further introduced a bright channel prior (BCP) to solve the corner case image, which contains a large amounts of bright pixels. By plugging the extreme channel prior (a combination of BCP and DCP) into the deblurring model, Yan et al. achieved stateoftheart results on various scenarios.
Although those algorithms demonstrate their effectiveness in image deblurring, the simplified assumptions on the blur model and timeconsuming parametertuning strategy are two lethal problems to hinder their performance in realworld cases. In this work, we utilize a realistic GoPro dataset [27] to endtoend train a new multiscale network for latent sharp image restoration.
2.2 Deep Learning based Deblurring Methods
Deep learning based methods focus on exploiting external training data to learn a mapping function accords with the degradation process. The powerful endtoend training paradigm and nonlinear modeling capability make CNNs a promising approach to image deblurring. Early CNNbased deblurring methods aim to mimic conventional deblurring frameworks for the estimation of both latent image and blur kernel. Works in [36] and [8] first used networks to predict the nonuniform blur kernel and then utilized a nonblind deblurring method [45] to restore images. Schuler et al. [32] introduced a twostages network to simulate iterative optimization. Chakrabarti et al. [3] utilized a network to predict frequency coefficients of blur kernel. However, these methods may fail when the estimated kernel is inaccurate [31]. Therefore, more recent approaches preferred to train kernel estimationfree networks to restore latent images directly. Specifically, Nah et al. [27] proposed a multiscale based CNN to progressively recover the latent image. Tao et al. [38] introduced a scalerecurrent network which equipped with a ConvLSTM layer [39] to further ensure information flow between different resolution images. Kupyn et al. [22] adopted Wasserstein GAN [9, 1] as an objective function to restore the texture details of latent image. Zhang et al. [43]
employed spatially variant recurrent neural networks (RNNs) to reduce the computational cost.
While existing deblurring networks have reported impressive results, the limited number of training data and the disappreciation of prior knowledge may become two main factors hampering the performance improvement. To mitigate aforementioned issues, we in this paper introduce the extreme channel prior for CNN and encourage it to constrain the solution space under these priors.
3 Extreme Channel Prior Embedded Network
Given a single blurry image , existing CNNbased methods aim at learning a mapping to generate an estimation of latent sharp image , which is required to approximate the groundtruth . This procedure can be formulated as:
(1) 
where (, ) refer to the th image pairs in the training dataset and is the parameter of mapping function.
However, such formulation is limited in capturing image priors specified to blind deblurring, which is generally much different from those for nonblind restoration. Moreover, the existing training image pairs are insufficient to learn an effective mapping function . Therefore, we propose ECPeNet that aggregates both blurry image and extreme channel representations to enhance deblurring performance, and Eqn. (1) can be rewritten as:
(2) 
where and are the extreme channel representations under the constraint of both dark and bright channels priors. By this way, we can embed the image priors specified to blind deblurring into the mapping function , and expect it to have ability to generate the higher quality latent sharp image. In the following subsections, we present in detail the proposed ECPeNet.
3.1 Architecture
The overall architecture of our proposed ECPeNet is illustrated in Figure 2
. It contains three subnetworks respectively for three scales, and each of them consists of three major components: (i) input and output; (ii) encoder and decoder; (iii) feature mapper. Note that instead of utilizing Rectifier Linear Units (ReLU)
[21], we take parametric ReLU (PReLU) [11]as activation function since it can improve the modeling capability with negligible extra computational cost. Unless denoted otherwise, all the convolution filters are set to
, not that utilized by most of the other dynamic scene deblurring networks (e.g., [27] and [38]). Although a filter of size has more parameters than two filters of size , the one utilizing filters is more efficient since additional nonlinearity can be inserted between them [35]. Besides, the stride size for all convolution layers is set to
and the number of feature maps in each layer is set to , except for the last layer and the proposed ECPeL which are set to and , respectively. The details of each component are described as follows.Input and Output. An effective multiscale network architecture is utilized in this work to restore the latent sharp image from a coarser scale to finer scale. Consider an image pair , where is the pixel resolution and is the number of channels, which is set to
. We first use bicubic interpolation to progressively downsample the image pair with the ratio of
, and generate scales image pairs with the resolution of . We then take the blurred image at each scales as input to produce its corresponding sharp images. The sharp one at original resolution is considered as the final output.Encoder and Decoder. Each scale of encoder consists of convolution layers, the proposed ECPeL and a shuffle operation with factor . As for decoders, they basically mirror the architecture of the encoders, except for the factor of shuffle operation [34] is set to . The encoder and decoder networks are mainly designed for three purposes.
Firstly, they progressively transform different scale images to extract shallow features (in encoders) and transform them back to the resolution of inputs (in decoders).
Secondly, they downsample and upsample the shallow features to ensure information flow between different scale images and expand the receptive field. Note that instead of directly concatenating the upsampled coarser scale latent image with the finer scale blurry image [27, 5, 38], we argue that (i) the information of finer scale blurry image is beneficial for estimating the coarser scale latent image and (ii) the concatenation in shallow feature domain can yield a better result than in image (RGB) domain. Thus, the encoder first shuffles features with factor to ensure the same resolution between different scales features, e.g., a features with size is shuffled to . Then it concatenates the downsampled features with coarser scale features for restoring the coarser scale sharp image. On the contrary, the decoder shuffles features with size back to and concatenates them with finer scale features for predicting the finer scale sharp image. Benefiting from this multiscale architecture, the network can fully exploit different scale images to maximize the information flow between them, resulting in better performance. In Section 4.1, we conduct an ablation study to verify its effectiveness.
Thirdly, they integrate the extreme channel prior into the network via ECPeL. We provide more details of the proposed ECPeL in Section 3.2.
Feature Mapper. The feature mapper module, which aims to refine the shallow features progressively, is an essential part of the latent image restoration. One critical factor for reducing blur artifacts is the size of receptive field. To enlarge the receptive field, we (i) stack a set of convolutional layers to achieve a larger depth network and (ii) utilize the shuffle operations for downsampling and upsampling the features. Considering that a deeper neural network is more challenging to converge, we adopt the residual blocks to speedup the training procedure [12, 15]. Besides, the feature mapper module utilizes the long skip connection and short skip connection to make full use of hierarchical features in all convolutional layers. A similar skip connection strategy has been utilized in a very recent work [44]. As illustrated in Figure 2, the feature mapper contains residual in residual blocks (RIRBlock), and each of them has residual blocks (ResBlock). The ResBlock consists of convolution layers and a PReLU activation function. Since we utilize the filter of size , the total parameters of our ECPeNet are almost in the same magnitude as the previous methods. Note that all the weights across different scales feature mapper submodules are shared.
3.2 Extreme Channel Prior Embedded Layer
The proposed ECPeL is designed for aggregating both blurry image representation and extreme channel representation to regularize the solution space of CNN. Specifically, it first learns mapping functions , and to transform the feature map from previous layer into new feature maps, including a deeper layer transformed feature , a dark channel prior constrained feature , and a bright channel prior constrained feature . It then adopts a concatenation operation to concatenate those feature maps for the integration of blurry image representation and extreme channel representation. Formally, the proposed ECPeL can be expressed as:
(3)  
where denotes the concatenation of the feature maps, and the subscripts and denote that parameters and are optimized under the dark and bright channel prior constraint. To add extreme channel prior constraint into a network, the ECPeL utilizes (i) extractors to extract both dark and bright channel of features, and (ii) the regularization term to enforce sparsity in training.
The extractor is designed to extract the dark channel of via computing its minimum values in a local patch. Formally, its forwards function can be written as follows:
(4)  
The extractor aims to extract the bright channel of by calculating its maximum values in a local patch. And its forwards function can be formulated as:
(5)  
where is the index set of inputs in a subwindow centered at a pixel location , and are the masks that records an index of the minimum and maximum value in a local patch, respectively. The patch sizes for each scale are set to . A single element and of the input may be assigned to several different outputs and .
The backwards
function of extractors computes partial derivative of the loss function with respect to each input variable
and as follows:(6)  
where refers to the pixel location . In words, the partial derivatives and are accumulated if is the and selected for and , respectively. In backpropagation, the partial derivatives and are already calculated by the backwards function of the loss layer.
With the proposed ECPeL, we can extract the dark and bright channel of shallow features (i.e., and ), which can be further enforced to be sparse via the objective function. By integrating the constrained features and into the network, the proposed ECPeNet can achieve a better performance while using the same training dataset. The ablation study in Section 4.1 is conducted for the evaluation.
3.3 Loss Function
We utilize the norm of the reconstruction error as loss function for each scale. More specifically, we can rewrite Eqn. (2) as:
(7) 
where is the total number of training pairs (, ) and is the number of scales, which is set to in this paper. is a symbol referring to the image and feature in the th scale.
According to the observation by [30, 42], a dark channel of images would be less dark after the blurring process, while a bright channel of images would be no longer bright. The reason is that dark/bright pixel would be averaged with its neighboring high/low intensity pixels during a blurring process. The sparsity regularization term is thus more beneficial for restoring a sharp image than a blurred one. To this end, we introduce a regularization term to enforce sparsity on both dark and bright channels of shallow features. The objective function can be given by:
(8)  
where and are the tradeoff parameters. and are the extractors to extract the dark channel and bright channel of features, respectively. With the forwards and backwards functions, the dark and bright channel extractors can be jointly endtoend optimized with the network.
4 Experimental Results
In this section, we provide experimental results to show the advantage of our proposed ECPeNet. We implement our framework by using Caffe toolbox
[14], and train the model on a PC equipped with an Intel Core i77820X CPU, 128G RAM and a single Nvidia Quadro GV100 GPU.Datasets. We train our proposed ECPeNet on the GoPro training dataset [27], which contains sequences with blurred/clear image pairs. Once the model is trained, we test it on the standard GoPro testing dataset [27] and Köhler [18] dataset. The GoPro testing dataset consists of sequences with image pairs, and the Köhler dataset has 4 latent images and 12 blur kernels. Note that, to simulate the realistic blurring process, the GoPro dataset generates blurred images through averaging adjacent shortexposure frames captured by a highspeed video camera, and the Köhler dataset replays the recorded real camera motion trajectory to synthesize blurred images.
Parameter Settings. We crop the GoPro training dataset (linear subset) into patches and make use of these patches to train the ECPeNet. The minibatch size in all the experiments is set to , and the tradeoff parameters and are set to in this paper. For the model training, we utilize Xavier [7] to initialize all trainable variables. The Adam solver [17] is adopted to optimize the network parameters. The default parameters of Adam solver are set as , and . We fix the learning rate as and train the network with iterations, which takes about hours. Additionally, we randomly rotate and/or flip the image patches for data augmentation. The additive Gaussian noise is also randomly added to the blurred images for robust learning.
Different combinations of ECP and IFE  
ECP  
IFE  
PSNR 
4.1 Ablation Study
It is generally agreed that a larger scale training dataset which covers various image contents and blur models will bring benefit to train a robust deep network. The type of scenes and number of images in the current GoPro dataset, however, are barely sufficient to train an efficient network. Rather than enlarging the training dataset, in this work, we propose to integrate the extreme channel prior into CNN and fully exploit different scales images for the performance improvement. Here, we compare our network with several baseline models to verify the effectiveness of the extreme channel prior (ECP) constraint and image fully exploitation (IFE) strategy. Table 1 shows the investigation on the effects of ECP constraint and IFE strategy. It can be seen that the one without both ECP and IFE performs much worse than the ECPeNet in terms of PSNR ( dB dB). By plugging the ECP constrain into the network, we can significantly improve the performance with a dB gain. While compared to the networks that utilizes the multiscale architectures [27, 38] or cascaded refinement [5] strategy, the one adopting the proposed IFE strategy (both coarsetofine and finetocoarse manners) can has dB improvement in terms of PSNR index. Note that we train all these networks with iterations for the testing in this ablation study. These comparisons firmly indicate the proposed ECP and IFE benefit for performance improvement.
Method  GoPro  Köhler  Time  
PSNR  SSIM  PSNR  MSSIM  
Kim [16]  23.64  0.824  24.68  0.794  hr 
Sun [36]  24.64  0.843  25.22  0.774  min 
Nah [27]  29.08  0.914  26.48  0.808  
Tao [38]  30.26  0.934  26.75  0.837  
Kupyn [22]  28.70  0.858  26.10  0.816  
Zhang [43]  29.19  0.931  25.71  0.800  
Proposed  31.10  0.945  26.79  0.839 
4.2 Comparisons with Stateoftheart Methods
In this subsection, both quantitative and qualitative evaluations are conducted to verify the proposed ECPeNet on the benchmark datasets.
Quantitative Evaluations. We first compare the proposed ECPeNet with previous stateoftheart deblurring methods [16, 36, 27, 22, 43, 38] in a quantitative way. The source codes and trained models of the aforementioned methods are publicly available on the authors’ websites, except for [16] and [43] whose results have been reported in previous works [27] and [43], respectively. Additionally, we utilize the same training dataset to retrain the network provided by [43] for its evaluation on Köhler dataset. The average PSNR, SSIM, and MSSIM indices for different deblurring methods on GoPro testing and Köhler datasets are shown in Table 2. One can see that on the GoPro testing dataset, the proposed ECPeNet significantly outperforms both the conventional nonuniform deblurring method [16] and these recent developed CNN based methods [36, 27, 22, 43, 38]. Even compared to the previous stateoftheart method [38], the proposed ECPeNet still has a dB lead. While on the Köhler dataset, although the performance of these dynamic scene deblurring networks is comparable, it can be found that our method still has a slight advantage. Meanwhile, the running time by different methods for processing an image of resolution is also listed in Table 2. One can notice that it takes plenty of time for a conventional method to restore an image because of the timeconsuming iterative inference and the CPU implementation. While for these endtoend training networks, they can achieve much faster speed to process an image on GPU. Considering that these dynamic scene deblurring networks are implemented by different deep learning platforms, the minor difference between them can be neglected within the margin of error.
Qualitative Evaluations. We further compare the visual quality of restored images by our proposed ECPeNet and these recent developed CNN based dynamic scene deblurring networks, including Nah [27], Tao [38], Kupyn [22], and Zhang [43]. Figure 3 shows several blurred images from the GoPro [27] testing dataset and their corresponding deblurring results produced by the above methods. One can see that although these recent developed CNNs could remove the overall motion blur artifacts, the results restored by them are not pleasant enough because of the blurred edges and noticeable artifacts. For example, in the fourth column, all these previous CNN based deblurring networks could not recover the text information (see the red box zoomin region). While in the second column, the noticeable artifacts exist around the license plate number. By contrast, benefiting from the extreme channel prior constraint, our method can deliver a more visual pleasing result with much fewer artifacts and sharper edges.
To further demonstrate the robustness of our method, the visual comparison results on images from the Köhler [18] dataset are also provided in Figure 4. Again, it can be seen that artifacts and blurred edges in the zoomin areas (see characters ‘’, ‘’, and ‘’) are noticeable for these previous CNN based methods. Although results recovered by Kupyn [22] and Tao [43] are sharper than other methods, distortion still exists. Compared with these methods, the ECPeNet can restore image sharpness and naturalness.
5 Conclusion
In this work, we presented a simple yet effective Extreme Channel Prior embedded Network (ECPeNet) with a novel trainable extreme channel prior embedded layer (ECPeL), which aims to integrate extreme (i.e., dark and bright) channel priors into a deep CNN for dynamic scene deblurring. By extracting the extreme channels of shallow features and enforcing sparsity on them, ECPeNet can regularize the solution space of the network. Additionally, ECPeNet works in both coarsetofine and finetocoarse manners to exploit information of blurred images at different resolutions to maximize information flow across scales. Benefiting from the extreme channel prior constraint and effective multiscale network architecture, the developed ECPeNet outperforms previous dynamic scene deblurring networks by a large margin. Quantitative evaluations on the challenging GoPro dataset showed that the proposed ECPeNet had at least dB PSNR gains over the existing stateofthearts.
References
 [1] M. Arjovsky, S. Chintala, and L. Bottou. Wasserstein gan. stat, 1050:26, 2017.
 [2] Y. Bahat, N. Efrat, and M. Irani. Nonuniform blind deblurring by reblurring. In ICCV, 2017.
 [3] A. Chakrabarti. A neural approach to blind motion deblurring. In ECCV, 2016.
 [4] T. F. Chan and C.K. Wong. Total variation blind deconvolution. IEEE TIP, 7(3):370–375, 1998.
 [5] Q. Chen and V. Koltun. Photographic image synthesis with cascaded refinement networks. In ICCV, 2017.
 [6] R. Fergus, B. Singh, A. Hertzmann, S. T. Roweis, and W. T. Freeman. Removing camera shake from a single photograph. In ACM transactions on graphics (TOG), volume 25, pages 787–794. ACM, 2006.
 [7] X. Glorot and Y. Bengio. Understanding the difficulty of training deep feedforward neural networks. In AISTATS, 2010.
 [8] D. Gong, J. Yang, L. Liu, Y. Zhang, I. Reid, C. Shen, A. Van Den Hengel, and Q. Shi. From motion blur to motion flow: A deep learning solution for removing heterogeneous motion blur. In CVPR, 2017.
 [9] I. Goodfellow, J. PougetAbadie, M. Mirza, B. Xu, D. WardeFarley, S. Ozair, A. Courville, and Y. Bengio. Generative adversarial nets. In NIPs, 2014.
 [10] K. He, J. Sun, and X. Tang. Single image haze removal using dark channel prior. CVPR, 2011.

[11]
K. He, X. Zhang, S. Ren, and J. Sun.
Delving deep into rectifiers: Surpassing humanlevel performance on imagenet classification.
In ICCV, 2015.  [12] K. He, X. Zhang, S. Ren, and J. Sun. Deep residual learning for image recognition. In CVPR, 2016.
 [13] Z. Hu, S. Cho, J. Wang, and M.H. Yang. Deblurring lowlight images with light streaks. In CVPR, 2014.
 [14] Y. Jia, E. Shelhamer, J. Donahue, S. Karayev, J. Long, R. Girshick, S. Guadarrama, and T. Darrell. Caffe: Convolutional architecture for fast feature embedding. In ACM MM, 2014.

[15]
J. Kim, J. K. Lee, and K. M. Lee.
Accurate image superresolution using very deep convolutional networks.
In CVPR, 2016.  [16] T. H. Kim and K. M. Lee. Segmentationfree dynamic scene deblurring. In CVPR, 2014.
 [17] D. P. Kingma and J. Ba. Adam: A method for stochastic optimization. In ICLR, 2014.
 [18] R. Köhler, M. Hirsch, B. Mohler, B. Schölkopf, and S. Harmeling. Recording and playback of camera shake: Benchmarking blind deconvolution with a realworld database. In ECCV, 2012.
 [19] D. Krishnan and R. Fergus. Fast image deconvolution using hyperlaplacian priors. In NIPs. 2009.
 [20] D. Krishnan, T. Tay, and R. Fergus. Blind deconvolution using a normalized sparsity measure. In CVPR, 2011.
 [21] A. Krizhevsky, I. Sutskever, and G. E. Hinton. Imagenet classification with deep convolutional neural networks. In NIPs, 2012.

[22]
O. Kupyn, V. Budzan, M. Mykhailych, D. Mishkin, and J. Matas.
Deblurgan: Blind motion deblurring using conditional adversarial networks.
In CVPR, 2018.  [23] A. Levin, Y. Weiss, F. Durand, and W. Freeman. Understanding and evaluating blind deconvolution algorithms. In CVPR, 2009.
 [24] L. Li, J. Pan, W.S. Lai, C. Gao, N. Sang, and M.H. Yang. Learning a discriminative prior for blind image deblurring. In CVPR, 2018.
 [25] L. B. Lucy. An iterative technique for the rectification of observed distributions. The astronomical journal, 79:745, 1974.
 [26] T. Michaeli and M. Irani. Blind deblurring using internal patch recurrence. In ECCV, 2014.
 [27] S. Nah, T. H. Kim, and K. M. Lee. Deep multiscale convolutional neural network for dynamic scene deblurring. In CVPR, 2017.
 [28] J. Pan, Z. Hu, Z. Su, and M.H. Yang. Deblurring face images with exemplars. In ECCV, 2014.
 [29] J. Pan, Z. Hu, Z. Su, and M.H. Yang. Lregularized intensity and gradient prior for deblurring text images and beyond. IEEE TPAMI, 39(2):342–355, 2017.
 [30] J. Pan, D. Sun, H. Pfister, and M.H. Yang. Blind image deblurring using dark channel prior. In CVPR, 2016.
 [31] D. Ren, W. Zuo, D. Zhang, J. Xu, and L. Zhang. Partial deconvolution with inaccurate blur kernel. IEEE TIP, 27(1):511–524, 2018.
 [32] C. J. Schuler, M. Hirsch, S. Harmeling, and B. Scholkopf. Learning to deblur. IEEE TPAMI, (7):1439–1451, 2016.
 [33] Q. Shan, J. Jia, and A. Agarwala. Highquality motion deblurring from a single image. In Acm transactions on graphics (TOG), volume 27, page 73. ACM, 2008.
 [34] W. Shi, J. Caballero, F. Huszár, J. Totz, A. P. Aitken, R. Bishop, D. Rueckert, and Z. Wang. Realtime single image and video superresolution using an efficient subpixel convolutional neural network. In CVPR, 2016.
 [35] K. Simonyan and A. Zisserman. Very deep convolutional networks for largescale image recognition. In ICLR, 2015.
 [36] J. Sun, W. Cao, Z. Xu, and J. Ponce. Learning a convolutional neural network for nonuniform motion blur removal. In CVPR, 2015.
 [37] L. Sun, S. Cho, J. Wang, and J. Hays. Edgebased blur kernel estimation using patch priors. In ICCP, 2013.
 [38] X. Tao, H. Gao, X. Shen, J. Wang, and J. Jia. Scalerecurrent network for deep image deblurring. In CVPR, 2018.

[39]
S. Xingjian, Z. Chen, H. Wang, D.Y. Yeung, W.K. Wong, and W.c. Woo.
Convolutional lstm network: A machine learning approach for precipitation nowcasting.
In NIPs, 2015.  [40] L. Xu and J. Jia. Twophase kernel estimation for robust motion deblurring. In ECCV, 2010.
 [41] L. Xu, S. Zheng, and J. Jia. Unnatural l sparse representation for natural image deblurring. In CVPR, 2013.
 [42] Y. Yan, W. Ren, Y. Guo, R. Wang, and X. Cao. Image deblurring via extreme channels prior. In CVPR, 2017.
 [43] J. Zhang, J. Pan, J. Ren, Y. Song, L. Bao, R. W. Lau, and M.H. Yang. Dynamic scene deblurring using spatially variant recurrent neural networks. In CVPR, 2018.
 [44] Y. Zhang, K. Li, K. Li, L. Wang, B. Zhong, and Y. Fu. Image superresolution using very deep residual channel attention networks. In ECCV, 2018.
 [45] D. Zoran and Y. Weiss. From learning models of natural image patches to whole image restoration. In ICCV, 2011.
 [46] W. Zuo, D. Ren, S. Gu, L. Lin, and L. Zhang. Discriminative learning of iterationwise priors for blind deconvolution. In CVPR, 2015.
Comments
There are no comments yet.