Single-frame Regularization for Temporally Stable CNNs

02/27/2019
by   Gabriel Eilertsen, et al.
2

Convolutional neural networks (CNNs) can model complicated non-linear relations between images. However, they are notoriously sensitive to small changes in the input. Most CNNs trained to describe image-to-image mappings generate temporally unstable results when applied to video sequences, leading to flickering artifacts and other inconsistencies over time. In order to use CNNs for video material, previous methods have relied on estimating dense frame-to-frame motion information (optical flow) in the training and/or the inference phase, or by exploring recurrent learning structures. We take a different approach to the problem, posing temporal stability as a regularization of the cost function. The regularization is formulated to account for different types of motion that can occur between frames, so that temporally stable CNNs can be trained without the need for video material or expensive motion estimation. The training can be performed as a fine-tuning operation, without architectural modifications of the CNN. Our evaluation shows that the training strategy leads to large improvements in temporal smoothness. Moreover, in situations where the quantity of training data is limited, the regularization can help in boosting the generalization performance to a much larger extent than what is possible with naïve augmentation strategies.

READ FULL TEXT

page 1

page 2

page 3

page 4

page 6

page 9

page 10

page 11

research
10/24/2021

AuxAdapt: Stable and Efficient Test-Time Adaptation for Temporally Consistent Video Semantic Segmentation

In video segmentation, generating temporally consistent results across f...
research
12/02/2021

Video Frame Interpolation without Temporal Priors

Video frame interpolation, which aims to synthesize non-exist intermedia...
research
01/11/2022

Motion-Focused Contrastive Learning of Video Representations

Motion, as the most distinct phenomenon in a video to involve the change...
research
01/15/2022

Learning Temporally and Semantically Consistent Unpaired Video-to-video Translation Through Pseudo-Supervision From Synthetic Optical Flow

Unpaired video-to-video translation aims to translate videos between a s...
research
02/21/2017

VidLoc: A Deep Spatio-Temporal Model for 6-DoF Video-Clip Relocalization

Machine learning techniques, namely convolutional neural networks (CNN) ...
research
03/31/2022

Time Lens++: Event-based Frame Interpolation with Parametric Non-linear Flow and Multi-scale Fusion

Recently, video frame interpolation using a combination of frame- and ev...

Please sign up or login with your details

Forgot password? Click here to reset