Dual-MTGAN: Stochastic and Deterministic Motion Transfer for Image-to-Video Synthesis

02/26/2021
by   Fu-En Yang, et al.
0

Generating videos with content and motion variations is a challenging task in computer vision. While the recent development of GAN allows video generation from latent representations, it is not easy to produce videos with particular content of motion patterns of interest. In this paper, we propose Dual Motion Transfer GAN (Dual-MTGAN), which takes image and video data as inputs while learning disentangled content and motion representations. Our Dual-MTGAN is able to perform deterministic motion transfer and stochastic motion generation. Based on a given image, the former preserves the input content and transfers motion patterns observed from another video sequence, and the latter directly produces videos with plausible yet diverse motion patterns based on the input image. The proposed model is trained in an end-to-end manner, without the need to utilize pre-defined motion features like pose or facial landmarks. Our quantitative and qualitative results would confirm the effectiveness and robustness of our model in addressing such conditioned image-to-video tasks.

READ FULL TEXT

page 1

page 5

page 6

research
07/17/2017

MoCoGAN: Decomposing Motion and Content for Video Generation

Visual signals in a video can be divided into content and motion. While ...
research
01/30/2021

Video Reenactment as Inductive Bias for Content-Motion Disentanglement

We introduce a self-supervised motion-transfer VAE model to disentangle ...
research
08/16/2023

Dual-Stream Diffusion Net for Text-to-Video Generation

With the emerging diffusion models, recently, text-to-video generation h...
research
12/17/2019

Jointly Trained Image and Video Generation using Residual Vectors

In this work, we propose a modeling technique for jointly training image...
research
11/05/2022

Disentangling Content and Motion for Text-Based Neural Video Manipulation

Giving machines the ability to imagine possible new objects or scenes fr...
research
06/14/2021

Flow Guided Transformable Bottleneck Networks for Motion Retargeting

Human motion retargeting aims to transfer the motion of one person in a ...
research
10/21/2019

DwNet: Dense warp-based network for pose-guided human video generation

Generation of realistic high-resolution videos of human subjects is a ch...

Please sign up or login with your details

Forgot password? Click here to reset