Learning to Extract a Video Sequence from a Single Motion-Blurred Image

by   Meiguang Jin, et al.
Universität Bern

We present a method to extract a video sequence from a single motion-blurred image. Motion-blurred images are the result of an averaging process, where instant frames are accumulated over time during the exposure of the sensor. Unfortunately, reversing this process is nontrivial. Firstly, averaging destroys the temporal ordering of the frames. Secondly, the recovery of a single frame is a blind deconvolution task, which is highly ill-posed. We present a deep learning scheme that gradually reconstructs a temporal ordering by sequentially extracting pairs of frames. Our main contribution is to introduce loss functions invariant to the temporal order. This lets a neural network choose during training what frame to output among the possible combinations. We also address the ill-posedness of deblurring by designing a network with a large receptive field and implemented via resampling to achieve a higher computational efficiency. Our proposed method can successfully retrieve sharp image sequences from a single motion blurred image and can generalize well on synthetic and real datasets captured with different cameras.


page 1

page 6

page 7

page 8


Affine-modeled video extraction from a single motion blurred image

A motion-blurred image is the temporal average of multiple sharp frames ...

Bringing Alive Blurred Moments!

We present a solution for the novel goal of extracting a video from a si...

Learning Blind Motion Deblurring

As handheld video cameras are now commonplace and available in every sma...

Unfolding a blurred image

We present a solution for the goal of extracting a video from a single m...

Self-supervised Exposure Trajectory Recovery for Dynamic Blur Estimation

Dynamic scene blurring is an important yet challenging topic. Recently, ...

Learning Event-Based Motion Deblurring

Recovering sharp video sequence from a motion-blurred image is highly il...

Stereo image de-fencing using smartphones

Conventional approaches to image de-fencing have limited themselves to u...

1 Introduction

It is often said that photos capture a memory, an instant in time. Technically, however, this is not strictly true. Photos require a finite exposure to accumulate light from the scene. Thus, objects moving during the exposure generate motion blur in a photo. Motion blur is an image degradation that makes visual content less interpretable and is therefore often seen as a nuisance. However, motion blur also combines information about both texture and motion of the objects in a single blurry image. Hence, recovering texture and motion from motion blurred images can be used to understand the dynamics of a scene (e.g., in entertainment with sports or in surveillance when monitoring the traffic). The task of recovering a blur kernel and a sharp image, whose convolution gives rise to a given blurry image, is called motion deblurring or blind deconvolution. Unfortunately, this formulation of the task is accurate only for some special cases of motion blur. In particular, it holds in the cases when blur is the same across an image (the so-called shift invariant blur [19]) or when blur can be modeled as a linear combination of a basis of shift fields (e.g., in the case of camera shake [5]). However, in the case of multiple moving objects, also called dynamic blur [21], a blurry image is no longer some convolution of a blur pattern with a single sharp image. In this case, a blurry image is the averaging over time of instant frames, where multiple objects move independently and cause occlusions.

In this paper we introduce blind deconvolution with dynamic blur as the task of recovering a sequence of sharp frames from a single blurry image. As illustrated in Fig. 1, given a single motion-blurred image (left column) we aim at recovering a sequence of

frames each depicting some instantaneous motion of the objects in the scene. To the best of our knowledge, this is the first time this problem has been posed and addressed. The two main challenges in solving this task are that: 1) blur removal is an ill-posed problem and 2) averaging over time destroys the temporal ordering of the instant frames. To handle the ill-posedness of deblurring we use a deep learning approach and train a convolutional neural network with a large receptive field. A large receptive field could be achieved by using large convolutional filters. However, such filters would have a detrimental impact on the memory requirements and the computational cost of the network. We avoid these issues by using a re-sampling layer (see Sec. 

6). Handling the loss of the temporal ordering is instead a less well-studied problem in the literature. To make matters worse, this ordering ambiguity extends to the motion of each object in the scene, thus leading to a combinatorial explosion of valid solutions. One possible exception to this scenario is the estimation of the frame in the middle of the sequence. In most motion blurred images the middle frame corresponds to the center of mass of the local blur, which can be unambiguously identified given the blurry input image [21, 23]. However, as we show in the Experiments section, the other frames do not enjoy the same uniqueness. We find that training a neural network by defining a loss on a specific frame of the sequence, other than the middle one, yields very poor results (see Sec. 7). We thus analyze temporal ambiguities in Sec. 4 and present a novel deep learning method that extracts instant frames in a sequential manner. Our main contribution is to train neural networks via loss functions that are invariant to the temporal ordering of the frames. These loss functions use the average of two frames and the absolute value of their difference as targets. This allows each network to choose which frames to output during training. Moreover, to make the network outputs more realistic and sharp, we use adversarial training [9]

. In the Experiments section we demonstrate that our trained networks can successfully extract videos from both synthetic and real motion-blurred images. In addition to providing accurate motion information about objects in the scene, we plan to use our method for video editing and temporal superresolution of videos. By exploiting the information embedded in motion blur, our method has the potential to interpolate subsequent frames with high accuracy.

2 Prior Work

Uniform Motion Deblurring. Blind deconvolution has been studied for several decades, and tremendous progress has been made in the case of uniform motion blur [26, 20, 2, 4, 35] and camera shake blur [10, 36, 1]. Yan  et al. [35] present an extremely effective image prior by combining the bright and dark channel priors of Pan et al[26]. Michaeli et al[20] incorporated recurrence of small image patches across different scales of a natural image. Gong  et al. [6] introduced a gradient activation algorithm for kernel estimation.

Non-Uniform Motion Deblurring. Recently, the general motion deblurring problem has attracted a lot of attention. [13] propose an energy model to estimate different motion blurs and their associated pixel-wise weights. [14] use a TV-L1 model to simultaneously estimate motion flow and a latent sharp image. Sun  et al. [33]

trained a convolutional neural network (CNN) for predicting a probability distribution of motion blurs. A sharp image is estimated by using a patch-level image prior. Pan

et al[25] developed an efficient algorithm to jointly estimate object segmentation and camera motion, where each layer is deblurred under the guidance of a soft-segmentation. [7] estimated a dense motion flow with a fully convolutional neural network and recovered the latent sharp image from the estimated motion flow. Bahat et al[1] recover the unknown blur field by analyzing the spectral content and deblur the image from the estimated blur field with a patch recurrence prior. Pan et al[24] proposed a method to learn data fitting functions from a large set of motion blurred images with the associated ground truth blur kernels. Nimisha et al[22] used adversarial training to learn blur-invariant features which fed to a decoder to produce a deblurred image. Recent work [21, 23, 15] suggests to generate synthetic data for dynamic scene motion blur by averaging consecutive frames captured with a high frame rate camera. This dataset could then be used to train a neural network. For example, [21] trained an end-to-end model with a multi-scale convolutional neural network to restore the latent image directly.

Video Deblurring. A number of methods consider the task of restoring a sharp sequence from a blurry video sequence. Zhang et al[37] propose a video deblurring method that jointly estimates the motion between consecutive frames as well as blur within each frame. Sellent et al[30] instead exploit a stereo video sequence. Wieschollek et al[34] introduce a recurrent network architecture to deblur images by taking temporal information into account. Kim et al[15] also exploit a (spatio-temporal) recurrent network, but aim for real-time performance. Kim et al[18] propose a method for simultaneously removing general blurs and estimating optical flow from a video sequence. Ren et al[29] exploit semantic segmentation of each blurry frame to understand the scene contents and use different motion models for image regions to guide the optical flow estimation. Su et al[32] propose a CNN that deblurs videos by incorporating information accumulated across frames. Pan et al[27] propose a framework to jointly estimate the scene flow and deblur the image. Park et al[28]

develop a method for the joint estimation of camera pose, depth, deblurring, and super-resolution from a blurred image sequence.

However, none of these approaches solves the task of extracting a video sequence from a single motion-blurred image. In the following sections, we first illustrate the main challenges of our problem, then we introduce our novel loss functions and show how they address these challenges. The network design is introduced in Sec. 6 and tested on synthetic and real datasets in the Experiments section.

3 From Video to Image

An image captured with exposure can be written as


where is the camera response function, which relates the irradiance at the image plane to the measured image intensity, and is the instant image (irradiance) at time . We discretize the time axis into segments, and define a sequence of frames , with . Each frame corresponds to the integral of over a segment, i.e.,


Object motion introduces a relative shift (in pixels) of regions between subsequent instant images . Given the maximum shift that we are interested in handling, and by defining negligible blur as a shift of pixel, we can define the maximum number of time segments by setting . This choice only ensures that on average each frame will have no motion blur. However, motions with acceleration may cause blur larger than pixel in some frames.

The motion-blur model (1) thus far described is quite general, as regions can shift in an unconstrained way and subsequent instant images can introduce or remove texture (occlusions). Indeed, this model can handle the most general case of motion-blur, often called dynamic blur. This suggests a new formulation of motion deblurring with dynamic blur:

Given a motion-blurred image , recover the frames satisfying model (1).

As mentioned in the Introduction, the task of recovering a sharp image from a blurry one is already known to be highly ill-posed. In our formulation, however, the task is made even more challenging by the loss of frame ordering in the model (1). It may be possible to determine the local ordering of subsequent frames by exploiting temporal smoothness, however, there exist several ambiguities that we describe and discuss in the next section. For example, given it is impossible to know if the ordering of the original sequence was or , which corresponds to all objects moving forward or backward in time. Due to the complexity of our task, we adopt a data-driven approach. We build a dataset of blurry images with corresponding ground truth frames by exploiting high frame-rate videos as in recent methods [21, 23], and devise a novel training method with convolutional neural networks (see Sec. 5.3).

(a)    (b)    (c)    (d)    (e)    (f)    (g)    (h)    (i)    (l)

Figure 2: Temporal ordering ambiguities. In this toy example, we show two moving objects: a red and a green ball. Both are translating horizontally. Columns (a)-(e) show five video frames in four scenarios. Each of the four rows shows a plausible motion scenario of the two objects. Column (f) shows the blurry average of the first five columns. These averages are all identical, thus demonstrating that all four sequences are equally valid solutions. Column (g) shows the average of frame (b) and (d). Column (h) shows the average of frame (a) and (e). Column (i) shows the absolute difference of frame (b) and (d). Finally, column (h) shows the absolute difference of frame (a) and (e).

4 Unraveling Time

In our data-driven approach we define a dataset of input data (a blurry image) and target (a sequence of frames) pairs and then train a neural network to learn this mapping. However, the averaging of frames in model (1) destroys the temporal ordering of the sequence . This makes the recovery of the frames very challenging, because it is not possible to define the target uniquely. We might expect that local temporal ambiguities between subsequent frames can be resolved by learning the temporal smoothness (frames are more likely to form a sequence describing smooth motions). However, several other ambiguities still remain. For example, the global motion direction is valid both forward and backward in time. This directional ambiguity applies independently to each moving object in the scene so that all motion direction combinations are valid.

We illustrate these ambiguities in Fig. 2 with a toy example. We consider two moving objects: a red and a green ball both translating along the horizontal axis. The first columns show all frames () in the averaging model. Because there are objects, there are possible combinations of motion directions ( motions with the number of objects). These are shown in the rows of the figure. Column (f) shows that the corresponding average of the frames is the same identical motion-blurred image in all cases. Therefore, any of the target frames across the rows is a valid one, and it would be unfeasible for the network to learn to predict a specific choice for just one of these

cases. Indeed, as we show in the Experiments section, training a network to predict a single frame results in a network that predicts a blurry output that is the average of the possible choices. There is one exception to these ambiguities. The middle frame in an odd-numbered sequence does not change across the

cases. This explains why prior methods [21, 23] could successfully train a neural network to predict the middle frame.

To address the temporal ordering ambiguities we introduce novel loss functions. In the next section we explore different options and show how we arrived at our proposed loss function. These cases are also discussed and evaluated in the Experiments section.

5 From Image to Video

Our training data has been obtained from a GoPro Hero 5 and features videos at 240 frames per second. To obtain blurry frames at a standard real-time video rates (30 frames per second) we thus need to average frames. However, as we have shown in our previous section we can avoid ambiguities in the estimation of the middle frame by using an odd number of frames, and hence we use . However, our method can generalize to other choices of . We denote the neural network that predicts the frame with . Since the middle frame can be predicted directly without ambiguities, we train with the following loss


where is the perceptual loss [16]. For the perceptual loss, we use the relu2_2 and relu3_3 layers of vgg16 net [31]. All other losses in the sections below will focus on the other frames.

5.1 Globally Ordering-Invariant Loss

A first way to recover all other frames is to use a loss function based on the image formation model (1)


where we have defined . This loss does not suffer from ambiguities and lets the networks decide what frames to output. In practice, however, we find that it is too weak. This loss works well only when a blurry frame is generated by averaging no more than 3 frames. We find experimentally that with more averaging frames, the network does not converge well and may not generate a meaningful sequence.

5.2 Pairwise Ordering-Invariant Loss

Inspired by the previous observation, we notice that any pair of symmetric frames (about the middle frame) results in the same average and absolute differences. This choice is motivated by the observations made in the previous section and illustrated in Fig. 2. Columns (g) and (i) in Fig. 2 show the average and absolute difference respectively of columns (b) and (d). These combinations yield the same target frame regardless of the objects motion direction. Thus, we propose to use a loss made of two components, one based on the sum and the other based on the absolute difference between only two frames. We find experimentally that this scheme imposes a much stronger constraint. Based on these observations, for each pair of symmetric frames , we propose the following loss function


where and for . Notice that takes as inputs both the blurry image and the output of the other network . The reason for this additional input is so that the network can learn to generate a frame different from that of . Therefore, it needs to “know” what frame the network has chosen to generate. Compared with the loss function in eq. (4), this loss function is easier to optimize and converges better (see results in the Experiments section). We also find experimentally that we can further boost the performance of our networks by additionally feeding the middle frame prediction to each network. That is, we define and for .

5.3 Learning a Temporal Direction

Up to this point, each pair of networks and operates independently from the other pairs. This is not ideal, as it leaves a binary ambiguity in the temporal ordering of each pair: We do not know if and or and . Thus, after training, one needs to find a smooth temporal ordering of the outputs of these networks for each new input.

To avoid this additional task, we sequentially train our networks and use the outputs of the previous networks to determine the ordering for each data sample during training. This is needed only for frames away from the middle core with . Moreover, once the temporal ordering chosen by the middle core networks is known to the other networks, there is no need to feed other inputs. Hence, we define the non-core networks as and for . In practice, we find that and can share weights for . This opens up the possibility of designing a recurrent network to predict all non-core frames. We also use an adversarial loss to enhance the accuracy of the output of each network . Except for the network that generates the middle frame, all other networks use the adversarial loss during training. We summarize our training losses and procedure in Table 1.

Training procedure

  1. Let and minimize

  2. Let ,
    and minimize

  3. Let , , with and minimize

Table 1: Summary of networks, loss functions and training.

6 Implementation

Our middle frame prediction network employs a residual learning strategy like many recent image restoration networks [21, 17]

. It consists of three parts, feature extraction, feature refinement and feature fusion. Feature extraction is done by a resampling convolution with a resampling factor equal to 4. In the feature refinement part, we use 12 residual blocks

[11], where each one includes two and one convolution layers with a pre-activation structure [12]. To further increase the receptive field, we replace standard convolutional layers in the middle with 6 residual blocks with dilated convolutions. The feature extraction and refinement parts work on grayscale images, and three color-refined features are generated separately. The feature fusion part works on color images to compensate misalignments from the three separately-generated color-refined features. For the non middle frame prediction networks, a similar structure is applied. The differences are the feature extraction part, where features are extracted from multiple inputs separately and then concatenated, resampling factor, and number of channels. More details of these architectures can be found in the supplementary material.

7 Experiments

In this section, we perform a quantitative comparison of the middle frame prediction network with the state-of-the-art method [21]. For non middle frame predictions, we carry out a qualitative evaluation as there is no existing method predicting a video sequence from a single motion-blurred input. We show some examples of video reconstructions from real motion blurred images in Fig. 6. More examples and videos are available in the supplementary material. We validate our design through ablation studies of different loss functions.

Method [21] (dB) Our testset (dB) [21] (dB)
Nah [21] 30.52 28.19 28.48
Middle 32.20 29.02 26.98
Table 2: Comparison of the middle frame prediction networks.
Method P P P params
Nah [21] 2.43 3.52 4.80 12M
Middle 0.24 0.30 0.45 5M
Full 0.61 0.74 1.10 17M
Table 3: Execution time comparison between the state of the art single image dynamic scene deblurring network [21] and our model on three different resolutions on a Titan X GPU.
Nah [21]
Blurry crop
Nah [21] crop
Proposed crop
Su [32]
Blurry crop
Su [32] crop
Proposed crop
Figure 3: Middle frame prediction comparison. The first and fourth columns show the blurry inputs and cropped regions. (a)-(d) second and fifth columns: frame predictions from [21] network. (e)-(g) second and fifth columns: frame predictions from [32] network. (a)-(g) third and last columns: frame predictions from our proposed network. The first two rows have been generated synthetically through averaging. The third and fourth rows are real images captured with a DSLR camera. The last three rows are real images from [3] and [32].

Training Dataset and Implementation Details. Although there is a GoPro training set available from [21], containing 22 diverse scenes, we captured additional 20 scenes. In the training, we downsample the GoPro frames to 45 of their original size ( pixels) to suppress noise. Blurry frames are generated by averaging consecutive frames randomly cropped of size . For training we use about 15K samples. Data augmentation is applied to avoid overfitting by randomly shuffling color channels, rotating images, and adding

white gaussian noise. Networks are implemented using PyTorch and training is done with 2 GTX 1080 Ti GPUs. The batch-size of the middle frame prediction network and other networks are 32 and 24, respectively. Training at each stage takes 1 day, and all together the full network training takes 4 days.

Frame 1
Frame 2
Frame 3
Frame 4
Frame 5
Frame 6
Frame 7
Figure 4: Ablation study on real data with different loss functions.
Blurry crop
Frame 1
Frame 4
Frame 7
Figure 5: A synthetic example from [21] test image. (b)-(d) Frame 4 show our estimated middle frame, Nah’s estimate and the ground truth, respectively. As can be seen, the middle frame estimates from both our method and Nah’s are incorrect and affect the estimates of the other frames. Only when the ground truth middle frame is provided, the other frames can be estimated correctly.

Middle Frame Reconstruction. We take Nah’s [21] test set, which contains 11 different sequences, and generate 1700 blurry frames by averaging 7 consecutive frames. The same process is also applied to our own test set where 450 blurry images are generated. All blurry images are downsampled to as during training. Table 2 shows the quantitative results of Nah’s network and our proposed network on two datasets. On the last two datasets (the first two columns in the table), our network is consistently preforming better. This is because the motion blur in the data matches the motion blur observed by our network during training. In contrast, Nah’s network was trained with much more challenging data, where motion blur could be even larger. Thus, for a fairer comparison, we also evaluate our network on Nah’s original test images. These images are averaged by more than frames without any downsampling. In this case, Nah’s network is performing better, as our network has not learned to deal with such large motion blur. However, the performance loss is not too significant.

Some visual comparisons on both synthetic and real images are shown in Fig. 3. Fig. 3 (a) and (b) show two synthetic examples, one with an extremely large blur from Nah’s original test images and the other one with a moderate blur from our test set. It can be seen that although our method does not outperform Nah’s, it can give better visual results when blur is moderate. Two real examples captured with a DSLR (Nikon D7100) are also shown in rows (c) and (d). In practice, we find that if a network is trained with large blurs, it may not remove moderate blur to the same extent as networks trained with small blurs. As we will show later in the Experiments section, the accuracy of the middle frame prediction has a dramatic impact on the reconstruction of the other frames.

Another advantage of our network is computational efficiency. In Table 3 we show the execution time for 3 different resolutions and number of parameters used in Nah’s and our networks. It can be seen that our middle frame prediction network is approximately 10 times faster than Nah’s, but has half as many parameters. Additionally, in Fig. 3 we also compare to the state of the art video deblurring method [32]. Notice that in [32], they use 5 consecutive blurry frames to predict the middle sharp frame, whereas we predict the middle frame directly from only one blurry input. Three real results are shown in Fig. 3 (e), (f) and (g). It can be seen that, although our method suffers from some jpeg artifacts, it gives comparable results.

Resampling Factors. We also evaluate the different resampling factors for the middle frame estimation and find that resampling gives a better trade-off between accuracy and execution time.

Blurry Blurry Blurry Blurry Blurry Blurry Blurry Blurry Blurry Blurry Blurry Blurry Blurry Blurry Blurry Blurry Blurry Blurry Blurry Blurry Blurry Blurry Blurry Blurry Blurry
Frame 1
Frame 2
Frame 3
Frame 4
Frame 5
Frame 6
Frame 7
Figure 6: Examples with real images. All images are captured with a DSLR camera. Top row: an image with multiple moving objects. Notice that the background scene is static. Second row Frame 1-7: the reconstructed video (requires zooming). Rows 3 and 4: details of the scene. Notice that the reconstruction shows both cars moving left to right. This is not the true motion (it would correspond to one vehicle reversing on the street). Bottom row: a rotating ball. The network can correctly reconstruct a video with a complex motion field.

Independent Frame Reconstruction. A straightforward way of estimating a full sharp sequence is to replicate the training for each frame by minimizing the loss where is a network to predict . In this section we show that this scheme is not applicable beyond middle frames, due to the temporal ordering ambiguity. Fig. 4 (a) and (g) show the visual results with an independent frame reconstruction scheme. From the results we can see that the quality of the reconstructed frame worsens as the distance from the middle frame increases.

Global Frame Reconstruction. In this section we show that the global ordering-invariant loss is not a good option either. Fig. 4 (b) and (h) show the reconstructed 7 frames, where the middle frame is reconstructed independently with the loss and the other 6 frames are reconstructed jointly with the globally ordering-invariant loss . We can see that the non middle frame prediction network does not converge well and generates artifacts.

Pairwise Frame Reconstruction. Fig. 4 (c), (d), (i) and (j) show the reconstructions with the pairwise ordering-invariant loss . Rows (d) and (j) show the case where the middle frame prediction is also fed to the network, while the third row shows the case without the middle frame prediction. There are two main limitations of using : 1) One has to manually reorder non middle frame predictions; 2) Although feeding the middle frame prediction to the network gives better visual results than in the case without it, still both of these two schemes generate artifacts, especially for the frames temporally away from the middle frame.

Sequential Pairwise Frame Reconstruction. Fig. 4 (e) and (k) show the visual results with a sequential pair-wise reconstruction scheme. Notice that this scheme and the pairwise ordering scheme only differ at the 4 frame predictions , and . We can see that the sequential scheme generates fewer artifacts especially at frames and in Fig. 4 (k).

Teacher Forcing. We also explore the teacher forcing

method used to train recurrent neural networks

[8]. During training we substitute the middle frame prediction with the ground truth middle frame in step 2 and 3 of our full training procedure in Table 1. This strategy brings several benefits: 1) In practice, we observe that the teacher forcing training strategy converges faster than with a standard sequential pairwise training; 2) It also gives visually better predictions as shown in Fig. 4 (l), where non middle frames are predicted by a network trained with teacher forcing; 3) The middle frame and non middle frame prediction training can be done in parallel. We use teacher forcing training as our default network training scheme.

We use 4 different networks to predict all frames: one to predict the middle frame, and the other three for middle-symmetric pair-wise frames. We found that sharing the parameters of the pair-wise networks for frames 1, 2, 6 and 7 during training would not result in a loss of visual accuracy. This could make predicting more than 7 frames feasible.

Importance of the Middle Frame Estimate. We observe experimentally that a good initialization is key in making the non middle frame prediction network work well. Fig. 5 (a) shows a blurry image and an enlarged detail with significant motion blur. In Fig. 5 (b) we show the reconstructions of frame 1, 4 (middle) and 7 with our trained network, where we used our estimated frame 4 to recover the other frames. In Fig. 5 (c) we show the corresponding frames reconstructed when feeding our network with Nah’s [21] frame 4 estimate. Both cases fail to reconstruct the middle frames and the other frames. However, when we feed our networks with the ground truth middle frame (see Fig. 5 (d)), they are able to correctly reconstruct the other frames.

Discussion. Although our system can predict 7 frames from a motion-blurred image, there are two main limitations. One main limitation of our approach is that it is not robust to large blurs. Whenever our middle frame prediction network fails to remove blur, the non middle frame prediction networks also fail.

8 Conclusions

In this paper we have presented a first method to reconstruct a video from a single motion-blurred image. We have shown that the task is more ambiguous than deblurring a single frame because the temporal ordering is lost in the motion-blurred image. We have presented a data-driven solution that allows a convolutional neural network choose a temporal ordering at the output. We have demonstrated our model on several datasets and have shown that it generalizes on real images captured with different cameras from those used to collect the training set.

Acknowledgements. MJ, GM, and PF acknowledge support from the Swiss National Science Foundation on projects 200021_153324 and 200021_165845.


  • [1] Y. Bahat, N. Efrat, and M. Irani. Non-uniform blind deblurring by reblurring. In ICCV, 2017.
  • [2] A. Chakrabarti. A neural approach to blind motion deblurring. In ECCV, 2016.
  • [3] S. Cho, J. Wang, and S. Lee. Video deblurring for hand-held cameras using patch-based synthesis. ACM Trans. Graph., 2012.
  • [4] J. Dong, J. Pan, Z. Su, and M.-H. Yang.

    Blind image deblurring with outlier handling.

    In ICCV, 2017.
  • [5] R. Fergus, B. Singh, A. Hertzmann, S. T. Roweis, and W. T. Freeman. Removing camera shake from a single photograph. In SIGGRAPH, 2006.
  • [6] D. Gong, M. Tan, Y. Zhang, A. van den Hengel, and Q. Shi. Blind image deconvolution by automatic gradient activation. In CVPR, 2016.
  • [7] D. Gong, J. Yang, L. Liu, Y. Zhang, I. Reid, C. Shen, A. van den Hengel, and Q. Shi. From motion blur to motion flow: A deep learning solution for removing heterogeneous motion blur. In CVPR, 2017.
  • [8] I. Goodfellow, Y. Bengio, and A. Courville. Deep Learning. MIT Press, 2016.
  • [9] I. Goodfellow, J. Pouget-Abadie, M. Mirza, B. Xu, D. Warde-Farley, S. Ozair, A. Courville, and Y. Bengio. Generative adversarial nets. In NIPS, 2014.
  • [10] S. Harmeling, B. Scholkopf, C. J. Schuler, and M. Hirsch. Fast removal of non-uniform camera shake. ICCV, 2011.
  • [11] K. He, X. Zhang, S. Ren, and J. Sun. Deep residual learning for image recognition. In CVPR, 2016.
  • [12] K. He, X. Zhang, S. Ren, and J. Sun. Identity mappings in deep residual networks. In ECCV, 2016.
  • [13] T. Hyun Kim, B. Ahn, and K. Mu Lee. Dynamic scene deblurring. In ICCV, 2013.
  • [14] T. Hyun Kim and K. Mu Lee. Segmentation-free dynamic scene deblurring. In CVPR, 2014.
  • [15] T. Hyun Kim, K. Mu Lee, B. Scholkopf, and M. Hirsch. Online video deblurring via dynamic temporal blending network. In ICCV, 2017.
  • [16] J. Johnson, A. Alahi, and L. Fei-Fei. Perceptual losses for real-time style transfer and super-resolution. In ECCV, 2016.
  • [17] J. Kim, J. Kwon Lee, and K. Mu Lee. Deeply-recursive convolutional network for image super-resolution. In CVPR, 2016.
  • [18] T. H. Kim, S. Nah, and K. M. Lee. Dynamic scene deblurring using a locally adaptive linear blur model. CoRR, 2016.
  • [19] W.-S. Lai, J.-B. Huang, Z. Hu, N. Ahuja, and M.-H. Yang. A comparative study for single image blind deblurring. In CVPR, 2016.
  • [20] T. Michaeli and M. Irani. Blind deblurring using internal patch recurrence. In ECCV, 2014.
  • [21] S. Nah, T. Hyun Kim, and K. Mu Lee. Deep multi-scale convolutional neural network for dynamic scene deblurring. In CVPR, 2017.
  • [22] T. M. Nimisha, A. Kumar Singh, and A. N. Rajagopalan. Blur-invariant deep learning for blind-deblurring. In ICCV, 2017.
  • [23] M. Noroozi, P. Chandramouli, and P. Favaro. Motion deblurring in the wild. In GCPR, 2017.
  • [24] J. Pan, J. Dong, Y.-W. Tai, Z. Su, and M.-H. Yang. Learning discriminative data fitting functions for blind image deblurring. In ICCV, 2017.
  • [25] J. Pan, Z. Hu, Z. Su, H.-Y. Lee, and M.-H. Yang. Soft-segmentation guided object motion deblurring. In CVPR, 2016.
  • [26] J. Pan, D. Sun, H. Pfister, and M.-H. Yang. Blind image deblurring using dark channel prior. In CVPR, 2016.
  • [27] L. Pan, Y. Dai, M. Liu, and F. Porikli. Simultaneous stereo video deblurring and scene flow estimation. In CVPR, 2017.
  • [28] H. Park and K. Mu Lee. Joint estimation of camera pose, depth, deblurring, and super-resolution from a blurred image sequence. In ICCV, 2017.
  • [29] W. Ren, J. Pan, X. Cao, and M.-H. Yang. Video deblurring via semantic segmentation and pixel-wise non-linear kernel. In ICCV, 2017.
  • [30] A. Sellent, C. Rother, and S. Roth. Stereo video deblurring. In ECCV, 2016.
  • [31] K. Simonyan and A. Zisserman. Very deep convolutional networks for large-scale image recognition. In ICLR, 2015.
  • [32] S. Su, M. Delbracio, J. Wang, G. Sapiro, W. Heidrich, and O. Wang. Deep video deblurring for hand-held cameras. In CVPR, 2017.
  • [33] J. Sun, W. Cao, Z. Xu, and J. Ponce. Learning a convolutional neural network for non-uniform motion blur removal. In CVPR, 2015.
  • [34] P. Wieschollek, M. Hirsch, B. Scholkopf, and H. P. A. Lensch. Learning blind motion deblurring. In ICCV, 2017.
  • [35] Y. Yan, W. Ren, Y. Guo, R. Wang, and X. Cao. Image deblurring via extreme channels prior. In CVPR, 2017.
  • [36] H. Zhang and D. P. Wipf. Non-uniform camera shake removal using a spatially-adaptive sparse penalty. In NIPS, 2013.
  • [37] H. Zhang and J. Yang. Intra-frame deblurring by leveraging inter-frame camera motion. In CVPR, 2015.