Occlusion Aware Unsupervised Learning of Optical Flow

11/16/2017 ∙ by Yang Wang, et al. ∙ University of Southern California Baidu, Inc. 0

It has been recently shown that a convolutional neural network can learn optical flow estimation with unsupervised learning. However, the performance of the unsupervised methods still has a relatively large gap compared to its supervised counterpart. Occlusion and large motion are some of the major factors that limit the current unsupervised learning of optical flow methods. In this work we introduce a new method which models occlusion explicitly and a new warping way that facilitates the learning of large motion. Our method shows promising results on Flying Chairs, MPI-Sintel and KITTI benchmark datasets. Especially on KITTI dataset where abundant unlabeled samples exist, our unsupervised method outperforms its counterpart trained with supervised learning.



There are no comments yet.


page 2

page 7

This week in AI

Get the week's most popular data science and artificial intelligence research sent straight to your inbox every Saturday.

1 Introduction

Video motion prediction, or namely optical flow, is a fundamental problem in computer vision. With the accurate optical flow prediction, one could estimate the 3D structure of a scene 

[18], segment moving objects based on motion cues [38], track objects in a complicated environment [11], and build important visual cues for many high level vision tasks such as video action recognition [45] and video object detection [60].

Traditionally, optical flow is formulated as a variational optimization problem with the goal of finding pixel correspondences between two consecutive video frames [23]. With the recent development of deep convolutional neural networks (CNNs) [32], deep learning based methods have been adopted to learn optical flow estimation, where the networks are either trained to compute discriminative image features for patch matching [21] or directly output the dense flow fields in an end-to-end manner [16]. One major advantage of the deep learning based methods compared to classical energy-based methods is the computational speed, where most state-of-the-art energy-based methods require 1-50 minutes to process a pair of images, while deep nets only need less than 100 milliseconds with a modern GPU.

Since most deep networks are built to predict flow using two consecutive frames and trained with supervised learning [26], it would require a large amount of training data to obtain reasonably high accuracy [35]. Unfortunately, most large-scale flow datasets are from synthetic movies and ground-truth motion labels in real world videos are generally hard to annotate [29]. To overcome this problem, unsupervised learning framework is proposed to utilize the resources of unlabeled videos  [30]. The overall strategy behind those unsupervised methods is that instead of directly training the neural nets with ground-truth flow, they use a photometric loss that measures the difference between the target image and the (inversely) warped subsequent image based on the dense flow field predicted from the fully convolutional networks. This allows the networks to be trained end-to-end with a large amount of unlabeled image pairs, overcoming the limitation from the lack of ground-truth flow annotations.

However, the performance of the unsupervised methods still has a relatively large gap compared to their supervised counterparts [41]. To further improve unsupervised flow estimation, we realize that occlusion and large motion are among the major factors that limit the current unsupervised learning methods. In this paper, we propose a new end-to-end deep neural architecture that carefully addresses these issues.

More specifically, the original baseline networks estimate motion and attempt to reconstruct every pixel in the target image. During reconstruction, there will be a fraction of pixels in the target image that have no source pixels due to occlusion. If we do not address this issue, it could limit the optical flow estimation accuracy since the loss function would prefer to compensate the occluded regions by moving other pixels. For example, in Fig. 

1, we would like to estimate the optical flow from frame 1 to frame 2, and reconstruct frame 1 by warping frame 2 with the estimated flow. Let us focus on the chair in the bottom left corner of the image. It moves in the down-left direction, and some part of the background is occluded by it. When we warp frame 2 back to frame 1 using the ground-truth flow (Fig. 1c), the resulting image (Fig. 1d) has two chairs in it. The chair on the top-right is the real chair, while the chair on the bottom-left is due to the occluded part of the background. Because the ground-truth flow of the background is zero, the chair in frame 2 is carried back to frame 1 to fill in the occluded background. Therefore, frame 2 warped by the ground-truth optical flow does not fully reconstruct frame 1. From the other perspective, if we use photometric loss of the entire image to guide the unsupervised learning of optical flow, the occluded area would not get the correct flow, which is illustrated in Fig. 1i. It has an extra chair in the flow trying to fill the occluded background with nearby pixels of similar appearance, and the corresponding warped image Fig. 1j has only one chair in it.

To address this issue, we explicitly allow the network to exploit the occlusion prediction caused by motion and incorporate it into the loss function. More concretely, we estimate the backward optical flow (Fig. 1g) and use it to generate the occlusion map for the warped frame (Fig. 1h). The white area in the occlusion map denotes the area in frame 1 that does not have a correspondence in frame 2. We train the network to only reconstruct the non-occluded area and do not penalize differences in the occluded area, so that the image warped by our estimated forward optical flow (Fig. 1e) can have two chairs in it (Fig. 1f) without incurring extra loss for the network.

Our work differs from previous unsupervised learning methods in four aspects. 1) We proposed a new end-to-end neural network that handles occlusion. 2) We developed a new warping method that can facilitate unsupervised learning of large motion. 3) We further improved the previous FlowNetS by introducing extra warped inputs during the decoder phase. 4) We introduced histogram equalization and channel representation that are useful for optical flow estimation. The last three components are created to mainly tackle the issue of large motion estimation.

As a result, our method significantly improves the unsupervised learning based optical flow estimation on multiple benchmark dataset including Flying Chairs, MPI-Sintel and KITTI. Our unsupervised networks even outperforms its supervised counterpart [16] on KITTI benchmark, where labeled data is limited compared to unlabeled data.

2 Related Work

Optical flow has been intensively studied in the past few decades [23, 34, 10, 49, 37]. Due to page limitation, we will briefly review the classical approaches and the recent deep learning approaches.

Optical flow estimation. Optical flow estimation was introduced as a fundamental computer vision problem since the pioneering works [23, 34]. Starting from then, the accuracy of optical flow estimation has been improving steadily as evidenced by the results on Middlebury [8] and MPI-Sintel [14] benchmark dataset. Most classical optical flow algorithms belong to the variants of the energy minimization problem with the brightness constancy and spatial smoothness assumptions [12, 42]. Other trends include a coarse-to-fine estimation or a hierarchical framework to deal with large motion [13, 55, 15, 6], a design of loss penalty to improve the robustness to lighting change and motion blur [59, 46, 22, 54], and a more sophisticated framework to handle occlusion [2, 50] which we will describe in more details in the next subsection.

Occlusion-aware optical flow estimation.

Since occlusion is a consequence of depth and motion, it is inevitable to model occlusion in order to accurately estimate flow. Most existing methods jointly estimate optical flow and occlusion. Based on the methodology, we divide them into three major groups. The first group treats occlusion as outliers and predict target pixels in the occluded regions as a constant value or through interpolation 

[47, 3, 4, 52]. The second group deals with occlusion by exploiting the symmetric property of optical flow and ignoring the loss penalty on predicted occluded regions  [51, 2, 25]. The last group builds more sophisticated frameworks such as modeling depth or a layered representation of objects to reason about occlusion [50, 48, 58, 43]. Our model is similar to the second group, such that we do not take account the difference where the occlusion happens into the loss function. To the best of our knowledge, we are the first to incorporate such kind of method with a neural network in an end-to-end trainable fashion. This helps our model to obtain more robust flow estimation around the occlusion boundary [27, 9].

Deep learning for optical flow. The success of deep learning innovates new optical flow models. [21] uses deep nets to extract discriminative features to compute optical flow through patch matching. [5] further extends the patch matching based methods by adding additional semantic information. Later, [7] proposes a robust thresholded hinge loss for Siamese networks to learn CNN-based patch matching features. [56] accelerates the processing of patch matching cost volume and obtains optical flow results with high accuracy and fast speed.

Meanwhile, [16, 26] propose FlowNet to directly compute dense flow prediction on every pixel through fully convolutional neural networks and train the networks with end-to-end supervised learning. [40] demonstrates that with a spatial pyramid network predicting in a coarse-to-fine fashion, a simple and small network can work quite accurately and efficiently on flow estimation. Later, [24] proposes a method for jointly estimating optical flow and temporally consistent semantic segmentation with CNN. The deep learning based methods obtain competitive accuracy across many benchmark optical flow datasets including MPI-Sintel [56] and KITTI [26] with a relatively faster computational speed. However, the supervised learning framework limits the extensibility of these works due to the lack of ground-truth flow annotation in other video datasets.

Unsupervised learning for optical flow. [39] first introduces an end-to-end differentiable neural architecture that allows unsupervised learning for video motion prediction and reports preliminary results on a weakly-supervised semantic segmentation task. Later, [30, 41, 1]

adopt a similar unsupervised learning architecture with a more detailed performance study on multiple optical flow benchmark datasets. A common philosophy behind these methods is that instead of directly supervising with ground-truth flow, these methods utilize the Spatial Transformer Networks 

[28] to warp the current images to produce a target image prediction and use photometric loss to guide back-propagation [17]. The whole framework can be further extended to estimate the depth, camera motion and optical flow simultaneously in an end-to-end manner [53]. This overcomes the flow annotation problem, but the flow estimation accuracy in previous works still lags behind the supervised learning methods. In this paper, we show that unsupervised learning can obtain competitive results to supervised learning models. After the initial submission of this paper, we became aware of a concurrent work [36] which tries to solve the occlusion problem in unsupervised optical flow learning with a symmetric-based approach.

Figure 2: Our network architecture. It contains two copies of FlowNetS[16] with shared parameters which estimates forward and backward optical flow respectively. The forward warping module generates an occlusion map from the backward flow. The backward warping module generates the warped image that is used to compare against the original frame 1 over the non-occluded area. There is also a smoothness term applied to the forward optical flow.

3 Network Structure and Method

We first give an overview of our network structure and then describe each of its components in details.

Overall structure. The schematic structure of our neural network is depicted in Fig. 2. Our network contains two copies of FlowNetS with shared parameters. The upper FlowNetS takes two stacked images ( and ) as input and outputs the forward optical flow () from to . The lower FlowNetS takes the reverse stacked images ( and ) as input and outputs the backward flow () from to .

The forward flow is used to warp to reconstruct through a Spatial Transformer Network similar to [30]. We call this backward warping, since the warping direction is different from the flow direction. The backward flow is used to generate the occlusion map () by forward warping. The occlusion map indicates the region in that is correspondingly occluded in (i.e. region in that does not have a correspondence in ).

The loss for training our network contains two parts: a photometric term () and a smoothness term (). For the photometric term, we compare the warped image and the original target image in the non-occluded region to obtain the photometric loss . Note that this is a key difference between our method and previous unsupervised learning methods. We also add a smoothness loss applied to to encourage a smooth flow solution.

Forward warping and occlusion map. We model the non-occluded region in as the range of  [2], which can be calculated with the following equation,

where is the range map value at location . are the image width and height, and are the horizontal and vertical components of .

Figure 3: Illustration of the forward warping module demonstrating how the occlusion map is generated using the backward optical flow. Here we only have horizontal component optical flow and where 1 denotes moving right, -1 denote moving left and 0 denotes stationary. In the occlusion map, 0 denotes occluded and 1 denotes non-occluded.

Since is continuous, the location of a pixel after being translated by a floating number might not be exactly on an image grid. We use reversed bilinear sampling to distribute the weight of the translated pixel to its nearest neighbors. The occlusion map can be obtained by simply thresholding the range map at the value of 1 and results in a soft map with value between 0 and 1. . The whole forward warping module is differentiable and can be trained end-to-end with the rest of the network.

In order to better illustrate the forward warping module, we provide a toy example in Fig. 3. and have only 4 pixels each, in which different letters represent different pixel values. The flow and reversed flow only have horizontal component which we show as and . The motion from to is that pixel A moves to the position of B and covers it, while pixel E in the background appears in . To calculate the occlusion map, we first create an image filled with ones and then translate them according to . Therefore, the one at the top-right corner is translated to the top-left corner leaving the top-right corner at the value of zero. The top-right corner (B) of is occluded by pixel A and can not find its corresponding pixel in which is consistent with the formulation we discussed above.

Figure 4: Illustration of the backward warping module with an enlarged search space. The large green box on the right side is a zoom view of the small green box on the left side.

Backward warping with a larger search space. The backward warping module is used to reconstruct from with forward optical flow . The method adopted here is similar to [30, 41] except that we include a larger search space. The problem with the original warping method is that the warped pixel value only depends on its four nearest neighbors, so if the target position is far away from the proposed position, the network will not get meaningful gradient signals. For example in Fig. 4, a particular pixel lands in the position of proposed by the estimated optical flow, and its value is a weighted sum of its four nearest neighbors. However, if the true optical flow land the pixel at , the network would not learn the correct gradient direction, and thus stuck at a local minimum. This problem is particularly severe in the case of large motion. Although one could use a multi-scale image pyramid to tackle the large motion problem, if the moving object is small or has a similar color to the background, the motion might not be visible in small scale images.

More concretely, when we use the estimated optical flow to warp back to reconstruct at a grid point , we first translate the grid point in (the yellow square) to in . Because the point is not on the grid point in , we need to do bilinear sampling to obtain its value. Normally, the value at is a weighted sum of its four nearest neighbors (black dots in the zoomed view on the right side of Fig. 4). We instead first search an enlarged neighbor (e.g. the blue dots at the outer circle in Fig. 4 together with the four nearest neightbors) around the point . For instance, if in the enlarged neighbor of point , the point that has the closest value to the target value is , we assign the value at the point to be a weighted sum of values at and three other symmetrical points (points labeled with red crosses in Fig. 4) with respect to point . By doing this, we can provide the neural network with gradient pointing towards the location of .

Loss term. The loss of our network contains two components: a photometric loss () and a smoothness loss (). We compute the photometric loss using the Charbonnier penalty formula over the non-occluded regions with both image brightness and image gradient.

where is the occlusion map defined in the above section, and , together indexes over pixel coordinates. The loss is normalized by the total non-occluded area size to prevent trivial solutions.

For the smoothness loss, we adopt an edge-aware formulation similar to [20], because motion boundaries usually coincide with image boundaries. Since the occluded area does not have a photometric loss, the optical flow estimation in the occluded area is solely guided by the smoothness loss. By using an edge-aware smoothness penalty, the optical flow in the occluded area would be similar to the values in its neighbor that has the closest appearance. We use both first-order and second-order derivatives of the optical flow in the smoothness loss term.

where controls the weight of edges, and indexes over partial derivative on and directions. The final loss is a weighted sum of the above four terms,

Figure 5:

Our modification to the FlowNetS structure at one of the decoding stage. On the left, we show the original FlowNetS structure. On the right, we show our modification of the FlowNetS structure. conv6 and conv5_1 are features extracted in the encoding phase and named after

[16]. Image1_6 and Image2_6 are input images downsampled 64 times. The decoding stages at other scales are modified accordingly.

Flow network details. Our inner flow network is adopted from FlowNetS [16]. Same as FlowNetS, we use a multi-scale scheme to guide the unsupervised learning by down-sampling images to different smaller scales. The only modification we made to the FlowNetS structure is that from coarser to finer scale during the refinement phase, we add the image warped by the coarser optical flow estimation and its corresponding photometric error map as extra inputs to estimate the finer scale optical flow in a fashion similar to FlowNet2 [26]. By doing this, each layer only needs to estimate the residual between the coarse and fine scale. The detailed network structure can be found in Fig. 5. Our modification only increases the number of parameters by 2% compared to the original FlownetS, and it moderately improves the result as seen in the later ablation study.

Preprocessing. In order to have better contrast for moving objects in the down-sampled images, we preprocess the image pairs by applying histogram equalization and augment the RGB image with a channel representation. The detailed channel representation can be found in [44]. We find both preprocessing steps improve the final optical flow estimation results.

3 Methods Chairs Sintel Clean Sintel Final KITTI 2012 KITTI 2015
test train test train test train test train test


FlowNetS [16] 2.71 4.50 7.42 5.45 8.43 8.26
FlowNetS+ft [16] (3.66) 6.96 (4.44) 7.76 7.52 9.1
SpyNet [40] 2.63 4.12 6.69 5.57 8.43 9.12
SpyNet+ft [40] (3.17) 6.64 (4.32) 8.36 8.25 10.1
FlowNet2 [26] 2.02 3.96 3.14 6.02 4.09 10.06
FlowNet2+ft [26] (1.45) 4.16 (2.01) 5.74 (1.28) 1.8 (2.3) 11.48%


DSTFlow [41] 5.11 6.93 10.40 7.82 11.11 16.98 24.30
DSTFlow-best [41] 5.11 (6.16) 10.41 (6.81) 11.27 10.43 12.4 16.79 39%
BackToBasic [30] 5.3 11.3 9.9
Ours 3.30 5.23 8.02 6.34 9.08 12.95 21.30
Ours+ft-Sintel 3.76 (4.03) 7.95 (5.95) 9.15 12.9 22.6
Ours-KITTI 7.41 7.92 3.55 4.2 8.88 31.2%
Table 1: Quantitative evaluation of our method on different benchmarks. The numbers reported here are all average end-point-error (EPE) except for the last column (KITTI2015 test) which is the percentage of erroneous pixels (Fl-all). A pixel is considered to be correctly estimated if the flow end-point error is <3px or <5%. The upper part of the table contains supervised methods and lower part of the table contains unsupervised methods. For all metrics, smaller is better. The best number for each category is highlighted in bold. The numbers in parentheses are results from network trained on the same set of data, and hence are not directly comparable to other results.

4 Experimental Results

We evaluate our methods on standard optical flow benchmark datasets including Flying Chairs [16], MPI-Sintel [14] and KITTI [19]

, and compare our results to existing deep learning based optical flow estimation (both supervised and unsupervised methods). We use the standard endpoint error (EPE) measure as the evaluation metric, which is the average Euclidean distance between the predicted flow and the ground truth flow over all pixels.

4.1 Implementation Details

Our network is trained end-to-end using Adam optimizer [31] with and . The learning rate is set to be for training from scratch and for fine-tuning. The experiments are performed on two Titan Z GPUs with a batch size of 8 or 16 depending on the input image resolution. The training converges after roughly a day. During training, we first assign equal weights to loss from different image scales and then progressively increase the weight on the larger scale image in a way similar to [35]. The hyper-parameters are set to be (1.0, 1.0, 10.0, 0.0, 10.0) for Flying Chairs and MPI-Sintel datasets, and (0.03, 3.0, 0.0, 10.0, 10.0) for KITTI dataset.Here we used higher weights of image gradient photometric loss and second-order smoothness loss for KITTI because the data has more lightning changes and its optical flow has more continuously varying intrinsic structure. In terms of data augmentaion, we only used horizontal flipping, vertical flipping and image pair order switching. During testing, our network only predicts forward flow, the total computational time on a Flying Chairs image pair is roughly 90 milliseconds with our Titian Z GPUs. Adding an extra 8 milliseconds for histogram equalization (an OpenCV CPU implementation), the total prediction time is around 100 milliseconds.

4.2 Quantitative and Qualitative Results

Figure 6: Qualitative examples for Sintel dataset. The top three rows are from Sintel Clean and the bottom three rows are from Sintel Final.
Figure 7: Qualitative examples for KITTI dataset. The top three rows are from KITTI 2012 and the bottom three rows are from KITTI 2015.
3 occlusion enlarged modified contrast Chairs Sintel Clean Sintel Final
handling search FlowNet enhancement test train train
5.11 6.93 7.82
4.51 6.80 7.32
4.27 6.49 7.11
4.14 6.38 7.08
4.62 6.60 7.33
4.04 6.09 7.04
3.76 5.70 6.54
3.30 5.23 6.34
Table 2: Ablation study

Table 1 summarizes the EPE of our method and previous state-of-the-art deep learning methods, including FlowNet [16], SpyNet [40], FlowNet2 [26], DSTFlow [41] and BackToBasic [30]. Because DSTFlow reported multiple variations of their results, we cite their best number across all of their results in ”DSTFlow-best” here.

Flying Chairs. Flying Chairs is a synthetic dataset created by superimposing images of chairs on background images from Flickr. It was originally created for training FlowNet in a supervised manner [16]. We use it to train our network without using any ground-truth flow. We randomly split the dataset into 95% training and 5% testing. We label this model as ”Ours” in Table 1. Our EPE is significantly smaller than the previous unsupervised methods (i.e. EPE decreases from 5.11 to 3.30) and is approaching the level of its corresponding supervised learning result (2.71).

MPI-Sintel. Since MPI-Sintel is relatively small and only contains around a thousand image pairs, we use the training data from both clean and final pass (without ground-truth) to fine-tune our network pretrained on Flying Chairs and the resulting model is labeled as ”Ours+ft-Sintel”. Compared to other unsupervised methods, we achieve a much better performance (e.g., EPE decreases from 10.40 to 7.95 on Sintel Clean test). Note that fine-tuning did not improve much here, largely due to the small number of training data. Fig.6 illustrates the qualitative result of our method on MPI-Sintel.

KITTI. The KITTI dataset is recorded under real-world driving conditions, and it has more unlabeled data than labeled data. Unsupervised learning methods would have an advantage in this scenario since they can learn from the large amount of unlabeled data. The training data we use here is similar to [41] which consists of multi-view extensions (20 frames for each sequence) from both KITTI2012 and KITTI2015. During training, we exclude two neighboring frames from the image pairs with ground-truth flow and testing pairs to avoid mixing training and testing data (i.e. not including frame number 9-12 in each multi-view sequence). We train the model from scratch since the optical flow in KITTI dataset has its own domain spatial structure (different from Flying Chairs) and abundant data. We label this model as ”Ours-KITTI” in Table 1.

Table 1 suggests that our method not only significantly outperforms existing unsupervised learning methods (i.e. improves EPE from 9.9 to 4.2 on KITTI 2012 test), but also outperforms its supervised counterpart (FlowNetS+ft) by a large margin, although there is still a gap compared to the state-of-the-art supervised network FlowNet2. Fig. 7 illustrates the qualitative results on KITTI. Our model correctly captures the occluded area caused by moving out of the frame. Our flow results are also free from the artifacts seen in DSTFlow (see [41] Figure 4c) in the occlusion area.

Occlusion Estimation. We also evaluate our occlusion estimation on MPI-Sintel and KITTI dataset which provide ground-truth occlusion labels between two consecutive frames. Among the literatures, we only find limited reports on occlusion estimation accuracy. Table 3 shows the occlusion estimation performance by calculating the maximum F-measure introduced in [33]. On MPI-Sintel, our method has a comparable result with previous non-neural-network based methods [33, 57]. On KITTI we obtain 0.95 and 0.88 for KITTI2012 and KITTI2015 respectively (we did not find published occlusion estimation result on KITTI). Note that S2D used ground-truth occlusion maps to do supervised training of their occlusion model.

3 Method Sintel Sintel KITTI KITTI
Clean Final 2012 2015
Our 0.54 0.48 0.95 0.88
S2D [33] 0.57
MODOF [57] 0.48
Table 3: Occlusion estimation evaluation. The numbers we present here is maximum F-measure. The S2D method is trained with ground-truth occlusion labels.

4.3 Ablation Study

We conduct systematic ablation analysis on different components added in our method. Table 2 shows the overall effects of them on Flying Chairs and MPI-Sintel. Our starting network is a FlowNetS without occlusion handling, which is the same configuration as [41].

Occlusion handling. The top two rows in Table 2 suggest that by only adding occlusion handling to the baseline network, the model improves its EPE from 5.11 to 4.51 on Flying-Chairs and from 7.82 to 7.32 on MPI-Sintel Final, which is significant.

Enlarged search. The effect of enlarged search is also significant. The bottom two rows in Table 2 show that adding enlarged search, the final EPE improves from 3.76 to 3.30 on Flying-Chairs and from 6.54 to 6.34 on MPI-Sintel Final.

Modified FlowNet. A small modification to the FlowNet also improves significantly, as suggested in the 5-th row in Table 2. By only adding a 2% more parameters and computation, the EPE improves from 5.11 to 4.62 on Flying-Chairs and from 7.82 to 7.33 on MPI-Sintel Final.

Contrast enhancement. We find that contrast enhancement is also a simple but very effective preprocessing step to improve the unsupervised optical flow learning. By comparing the 4th row and last row in Table 2, we find the final EPE improves from 4.14 to 3.30 on Flying-Chairs and 7.08 to 6.34 on MPI-Sintel Final.

Combining all components. We also find that sometimes one component is not significant by itself, but the overall model improves dramatically when we add all the 4 components into our framework.

Effect of data. We have tried to use more data from KITTI raw videos (60,000 samples compared to 25,000 samples used in the paper) to train our model, but we did not find any improvement. We have also tried to adopt the network structure from SpyNet [40] and train them using our unsupervised method. However we did not get better result either, which suggests that the learning capability of our model is still the limiting factor, although we have pushed this forward by a large margin.

5 Conclusion

We present a new end-to-end unsupervised learning framework for optical flow prediction. We show that with modeling occlusion and large motion, our unsupervised approach yields competitive results on multiple benchmark datasets. This is promising since it opens a new path for training neural networks to predict optical flow with a vast amount of unlabeled videos and apply the flow estimation for more higher level computer vision tasks.


  • [1] A. Ahmadi and I. Patras. Unsupervised convolutional neural networks for motion estimation. In Image Processing (ICIP), 2016 IEEE International Conference on, pages 1629–1633. IEEE, 2016.
  • [2] L. Alvarez, R. Deriche, T. Papadopoulo, and J. Sánchez. Symmetrical dense optical flow estimation with occlusions detection. International Journal of Computer Vision, 75(3):371–385, 2007.
  • [3] A. Ayvaci, M. Raptis, and S. Soatto. Occlusion detection and motion estimation with convex optimization. In Advances in neural information processing systems, pages 100–108, 2010.
  • [4] A. Ayvaci, M. Raptis, and S. Soatto. Sparse occlusion detection with optical flow. International Journal of Computer Vision, 97(3):322–338, 2012.
  • [5] M. Bai, W. Luo, K. Kundu, and R. Urtasun. Exploiting semantic information and deep matching for optical flow. In European Conference on Computer Vision, pages 154–170. Springer, 2016.
  • [6] C. Bailer, B. Taetz, and D. Stricker. Flow fields: Dense correspondence fields for highly accurate large displacement optical flow estimation. In Proceedings of the IEEE International Conference on Computer Vision, pages 4015–4023, 2015.
  • [7] C. Bailer, K. Varanasi, and D. Stricker. Cnn-based patch matching for optical flow with thresholded hinge embedding loss. In

    IEEE Conference on Computer Vision and Pattern Recognition (CVPR)

    , 2017.
  • [8] S. Baker, D. Scharstein, J. Lewis, S. Roth, M. J. Black, and R. Szeliski. A database and evaluation methodology for optical flow. International Journal of Computer Vision, 92(1):1–31, 2011.
  • [9] C. Ballester, L. Garrido, V. Lazcano, and V. Caselles. A tv-l1 optical flow method with occlusion detection. Pattern Recognition, pages 31–40, 2012.
  • [10] M. J. Black and P. Anandan. The robust estimation of multiple motions: Parametric and piecewise-smooth flow fields. Computer vision and image understanding, 63(1):75–104, 1996.
  • [11] J.-Y. Bouguet. Pyramidal implementation of the affine lucas kanade feature tracker description of the algorithm. Intel Corporation, 5(1-10):4, 2001.
  • [12] T. Brox, A. Bruhn, N. Papenberg, and J. Weickert. High accuracy optical flow estimation based on a theory for warping. Computer Vision-ECCV 2004, pages 25–36, 2004.
  • [13] T. Brox and J. Malik. Large displacement optical flow: descriptor matching in variational motion estimation. IEEE transactions on pattern analysis and machine intelligence, 33(3):500–513, 2011.
  • [14] D. J. Butler, J. Wulff, G. B. Stanley, and M. J. Black. A naturalistic open source movie for optical flow evaluation. In European Conference on Computer Vision, pages 611–625. Springer, 2012.
  • [15] Z. Chen, H. Jin, Z. Lin, S. Cohen, and Y. Wu. Large displacement optical flow from nearest neighbor fields. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 2443–2450, 2013.
  • [16] A. Dosovitskiy, P. Fischer, E. Ilg, P. Hausser, C. Hazirbas, V. Golkov, P. van der Smagt, D. Cremers, and T. Brox. Flownet: Learning optical flow with convolutional networks. In Proceedings of the IEEE International Conference on Computer Vision, pages 2758–2766, 2015.
  • [17] C. Finn, I. Goodfellow, and S. Levine. Unsupervised learning for physical interaction through video prediction. In Advances in Neural Information Processing Systems, pages 64–72, 2016.
  • [18] D. Forsyth and J. Ponce. Computer vision: a modern approach. Upper Saddle River, NJ; London: Prentice Hall, 2011.
  • [19] A. Geiger, P. Lenz, and R. Urtasun. Are we ready for autonomous driving? the kitti vision benchmark suite. In Computer Vision and Pattern Recognition (CVPR), 2012 IEEE Conference on, pages 3354–3361. IEEE, 2012.
  • [20] C. Godard, O. Mac Aodha, and G. J. Brostow. Unsupervised monocular depth estimation with left-right consistency. In CVPR, volume 2, page 7, 2017.
  • [21] F. Güney and A. Geiger. Deep discrete flow. In Asian Conference on Computer Vision, pages 207–224. Springer, 2016.
  • [22] D. Hafner, O. Demetz, and J. Weickert. Why is the census transform good for robust optic flow computation? In International Conference on Scale Space and Variational Methods in Computer Vision, pages 210–221. Springer, 2013.
  • [23] B. K. Horn and B. G. Schunck. Determining optical flow. Artificial intelligence, 17(1-3):185–203, 1981.
  • [24] J. Hur and S. Roth. Joint optical flow and temporally consistent semantic segmentation. In European Conference on Computer Vision, pages 163–177. Springer, 2016.
  • [25] J. Hur and S. Roth. Mirrorflow: Exploiting symmetries in joint optical flow and occlusion estimation. arXiv preprint arXiv:1708.05355, 2017.
  • [26] E. Ilg, N. Mayer, T. Saikia, M. Keuper, A. Dosovitskiy, and T. Brox. Flownet 2.0: Evolution of optical flow estimation with deep networks. In IEEE Conference on Computer Vision and Pattern Recognition (CVPR), volume 2, 2017.
  • [27] S. Ince and J. Konrad. Occlusion-aware optical flow estimation. IEEE Transactions on Image Processing, 17(8):1443–1451, 2008.
  • [28] M. Jaderberg, K. Simonyan, A. Zisserman, et al. Spatial transformer networks. In Advances in Neural Information Processing Systems, pages 2017–2025, 2015.
  • [29] J. Janai, F. Güney, J. Wulff, M. Black, and A. Geiger. Slow flow: Exploiting high-speed cameras for accurate and diverse optical flow reference data. In Conference on Computer Vision and Pattern Recognition (CVPR), 2017.
  • [30] J. Y. Jason, A. W. Harley, and K. G. Derpanis. Back to basics: Unsupervised learning of optical flow via brightness constancy and motion smoothness. In Computer Vision–ECCV 2016 Workshops, pages 3–10. Springer, 2016.
  • [31] D. Kingma and J. Ba. Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980, 2014.
  • [32] A. Krizhevsky, I. Sutskever, and G. E. Hinton. Imagenet classification with deep convolutional neural networks. In Advances in neural information processing systems, pages 1097–1105, 2012.
  • [33] M. Leordeanu, A. Zanfir, and C. Sminchisescu. Locally affine sparse-to-dense matching for motion and occlusion estimation. In Proceedings of the IEEE International Conference on Computer Vision, pages 1721–1728, 2013.
  • [34] B. D. Lucas, T. Kanade, et al. An iterative image registration technique with an application to stereo vision. 1981.
  • [35] N. Mayer, E. Ilg, P. Hausser, P. Fischer, D. Cremers, A. Dosovitskiy, and T. Brox. A large dataset to train convolutional networks for disparity, optical flow, and scene flow estimation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 4040–4048, 2016.
  • [36] S. Meister, J. Hur, and S. Roth. UnFlow: Unsupervised learning of optical flow with a bidirectional census loss. In AAAI, New Orleans, Louisiana, Feb. 2018.
  • [37] M. Menze, C. Heipke, and A. Geiger. Discrete optimization for optical flow. In German Conference on Pattern Recognition, pages 16–28. Springer, 2015.
  • [38] D. Pathak, R. Girshick, P. Dollár, T. Darrell, and B. Hariharan. Learning features by watching objects move. In Proc. CVPR, volume 2, 2017.
  • [39] V. Patraucean, A. Handa, and R. Cipolla. Spatio-temporal video autoencoder with differentiable memory. arXiv preprint arXiv:1511.06309, 2015.
  • [40] A. Ranjan and M. J. Black. Optical flow estimation using a spatial pyramid network. In IEEE Conference on Computer Vision and Pattern Recognition (CVPR), volume 2, 2017.
  • [41] Z. Ren, J. Yan, B. Ni, B. Liu, X. Yang, and H. Zha. Unsupervised deep learning for optical flow estimation. In AAAI, pages 1495–1501, 2017.
  • [42] J. Revaud, P. Weinzaepfel, Z. Harchaoui, and C. Schmid. Epicflow: Edge-preserving interpolation of correspondences for optical flow. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 1164–1172, 2015.
  • [43] L. Sevilla-Lara, D. Sun, V. Jampani, and M. J. Black. Optical flow with semantic segmentation and localized layers. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 3889–3898, 2016.
  • [44] L. Sevilla-Lara, D. Sun, E. G. Learned-Miller, and M. J. Black. Optical flow estimation with channel constancy. In European Conference on Computer Vision, pages 423–438. Springer, 2014.
  • [45] K. Simonyan and A. Zisserman. Two-stream convolutional networks for action recognition in videos. In Advances in neural information processing systems, pages 568–576, 2014.
  • [46] F. Stein. Efficient computation of optical flow using the census transform. In DAGM-symposium, volume 2004, pages 79–86. Springer, 2004.
  • [47] C. Strecha, R. Fransens, and L. J. Van Gool. A probabilistic approach to large displacement optical flow and occlusion detection. In ECCV Workshop SMVP, pages 71–82. Springer, 2004.
  • [48] D. Sun, C. Liu, and H. Pfister. Local layering for joint motion estimation and occlusion detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 1098–1105, 2014.
  • [49] D. Sun, S. Roth, and M. J. Black. Secrets of optical flow estimation and their principles. In Computer Vision and Pattern Recognition (CVPR), 2010 IEEE Conference on, pages 2432–2439. IEEE, 2010.
  • [50] D. Sun, E. B. Sudderth, and M. J. Black. Layered image motion with explicit occlusions, temporal consistency, and depth ordering. In Advances in Neural Information Processing Systems, pages 2226–2234, 2010.
  • [51] J. Sun, Y. Li, S. B. Kang, and H.-Y. Shum. Symmetric stereo matching for occlusion handling. In Computer Vision and Pattern Recognition, 2005. CVPR 2005. IEEE Computer Society Conference on, volume 2, pages 399–406. IEEE, 2005.
  • [52] M. Unger, M. Werlberger, T. Pock, and H. Bischof. Joint motion estimation and segmentation of complex scenes with label costs and occlusion modeling. In Computer Vision and Pattern Recognition (CVPR), 2012 IEEE Conference on, pages 1878–1885. IEEE, 2012.
  • [53] S. Vijayanarasimhan, S. Ricco, C. Schmid, R. Sukthankar, and K. Fragkiadaki. Sfm-net: Learning of structure and motion from video. arXiv preprint arXiv:1704.07804, 2017.
  • [54] C. Vogel, S. Roth, and K. Schindler. An evaluation of data costs for optical flow. In German Conference on Pattern Recognition, pages 343–353. Springer, 2013.
  • [55] P. Weinzaepfel, J. Revaud, Z. Harchaoui, and C. Schmid. Deepflow: Large displacement optical flow with deep matching. In Proceedings of the IEEE International Conference on Computer Vision, pages 1385–1392, 2013.
  • [56] J. Xu, R. Ranftl, and V. Koltun. Accurate optical flow via direct cost volume processing. In The IEEE Conference on Computer Vision and Pattern Recognition (CVPR), July 2017.
  • [57] L. Xu, J. Jia, and Y. Matsushita. Motion detail preserving optical flow estimation. IEEE Transactions on Pattern Analysis and Machine Intelligence, 34(9):1744–1757, 2012.
  • [58] K. Yamaguchi, D. McAllester, and R. Urtasun. Efficient joint segmentation, occlusion labeling, stereo and flow estimation. In European Conference on Computer Vision, pages 756–771. Springer, 2014.
  • [59] R. Zabih and J. Woodfill. Non-parametric local transforms for computing visual correspondence. In European conference on computer vision, pages 151–158. Springer, 1994.
  • [60] X. Zhu, Y. Wang, J. Dai, L. Yuan, and Y. Wei. Flow-guided feature aggregation for video object detection. arXiv preprint arXiv:1703.10025, 2017.