G^3AN: This video does not exist. Disentangling motion and appearance for video generation

12/11/2019
by   Yaohui Wang, et al.
51

Creating realistic human videos introduces the challenge of being able to simultaneously generate both appearance, as well as motion. To tackle this challenge, we propose the novel spatio-temporal GAN-architecture G^3AN, which seeks to capture the distribution of high dimensional video data and to model appearance and motion in disentangled manner. The latter is achieved by decomposing appearance and motion in a three-stream Generator, where the main stream aims to model spatio-temporal consistency, whereas the two auxiliary streams augment the main stream with multi-scale appearance and motion features, respectively. An extensive quantitative and qualitative analysis shows that our model systematically and significantly outperforms state-of-the-art methods on the facial expression datasets MUG and UvA-NEMO, as well as the Weizmann and UCF101 datasets on human action. Additional analysis on the learned latent representations confirms the successful decomposition of appearance and motion.

READ FULL TEXT

page 3

page 6

page 7

page 8

research
05/06/2023

LEO: Generative Latent Image Animator for Human Video Synthesis

Spatio-temporal coherency is a major challenge in synthesizing high qual...
research
12/03/2018

A Two-Stream Variational Adversarial Network for Video Generation

Video generation is an inherently challenging task, as it requires the m...
research
07/07/2018

Video Prediction with Appearance and Motion Conditions

Video prediction aims to generate realistic future frames by learning dy...
research
01/18/2022

Autoencoding Video Latents for Adversarial Video Generation

Given the three dimensional complexity of a video signal, training a rob...
research
08/30/2023

MMVP: Motion-Matrix-based Video Prediction

A central challenge of video prediction lies where the system has to rea...
research
01/08/2021

InMoDeGAN: Interpretable Motion Decomposition Generative Adversarial Network for Video Generation

In this work, we introduce an unconditional video generative model, InMo...
research
03/09/2020

Motion-Attentive Transition for Zero-Shot Video Object Segmentation

In this paper, we present a novel Motion-Attentive Transition Network (M...

Please sign up or login with your details

Forgot password? Click here to reset