Frame Difference-Based Temporal Loss for Video Stylization

02/11/2021
by   Jianjin Xu, et al.
1

Neural style transfer models have been used to stylize an ordinary video to specific styles. To ensure temporal inconsistency between the frames of the stylized video, a common approach is to estimate the optic flow of the pixels in the original video and make the generated pixels match the estimated optical flow. This is achieved by minimizing an optical flow-based (OFB) loss during model training. However, optical flow estimation is itself a challenging task, particularly in complex scenes. In addition, it incurs a high computational cost. We propose a much simpler temporal loss called the frame difference-based (FDB) loss to solve the temporal inconsistency problem. It is defined as the distance between the difference between the stylized frames and the difference between the original frames. The differences between the two frames are measured in both the pixel space and the feature space specified by the convolutional neural networks. A set of human behavior experiments involving 62 subjects with 25,600 votes showed that the performance of the proposed FDB loss matched that of the OFB loss. The performance was measured by subjective evaluation of stability and stylization quality of the generated videos on two typical video stylization models. The results suggest that the proposed FDB loss is a strong alternative to the commonly used OFB loss for video stylization.

READ FULL TEXT

page 2

page 4

page 5

page 7

page 8

page 9

page 10

research
07/10/2020

Optical Flow Distillation: Towards Efficient and Stable Video Style Transfer

Video style transfer techniques inspire many exciting applications on mo...
research
12/03/2020

Learning to Transfer Visual Effects from Videos to Images

We study the problem of animating images by transferring spatio-temporal...
research
03/04/2020

Occlusion Aware Unsupervised Learning of Optical Flow From Video

In this paper, we proposed an unsupervised learning method for estimatin...
research
04/20/2020

End-to-End Learning for Video Frame Compression with Self-Attention

One of the core components of conventional (i.e., non-learned) video cod...
research
10/08/2012

Video De-fencing

This paper describes and provides an initial solution to a novel video e...
research
06/08/2018

Convolutional Video Steganography with Temporal Residual Modeling

Steganography represents the art of unobtrusively concealing a secrete m...
research
02/12/2020

Patternless Adversarial Attacks on Video Recognition Networks

Deep neural networks for classification of videos, just like image class...

Please sign up or login with your details

Forgot password? Click here to reset