Continuous-Time Video Generation via Learning Motion Dynamics with Neural ODE

12/21/2021
by   Kangyeol Kim, et al.
8

In order to perform unconditional video generation, we must learn the distribution of the real-world videos. In an effort to synthesize high-quality videos, various studies attempted to learn a mapping function between noise and videos, including recent efforts to separate motion distribution and appearance distribution. Previous methods, however, learn motion dynamics in discretized, fixed-interval timesteps, which is contrary to the continuous nature of motion of a physical body. In this paper, we propose a novel video generation approach that learns separate distributions for motion and appearance, the former modeled by neural ODE to learn natural motion dynamics. Specifically, we employ a two-stage approach where the first stage converts a noise vector to a sequence of keypoints in arbitrary frame rates, and the second stage synthesizes videos based on the given keypoints sequence and the appearance noise vector. Our model not only quantitatively outperforms recent baselines for video generation, but also demonstrates versatile functionality such as dynamic frame rate manipulation and motion transfer between two datasets, thus opening new doors to diverse video generation applications.

READ FULL TEXT

page 2

page 7

page 9

page 10

page 20

page 21

page 22

page 23

research
10/16/2020

Vid-ODE: Continuous-Time Video Generation with Neural Ordinary Differential Equation

Video generation models often operate under the assumption of fixed fram...
research
04/12/2023

VidStyleODE: Disentangled Video Editing via StyleGAN and NeuralODEs

We propose VidStyleODE, a spatiotemporally continuous disentangled Video...
research
11/10/2021

Dance In the Wild: Monocular Human Animation with Neural Dynamic Appearance Synthesis

Synthesizing dynamic appearances of humans in motion plays a central rol...
research
08/11/2020

DTVNet: Dynamic Time-lapse Video Generation via Single Still Image

This paper presents a novel end-to-end dynamic time-lapse video generati...
research
06/06/2023

Learn the Force We Can: Multi-Object Video Generation from Pixel-Level Interactions

We propose a novel unsupervised method to autoregressively generate vide...
research
04/16/2023

CAT-NeRF: Constancy-Aware Tx^2Former for Dynamic Body Modeling

This paper addresses the problem of human rendering in the video with te...
research
04/08/2018

Learning-based Video Motion Magnification

Video motion magnification techniques allow us to see small motions prev...

Please sign up or login with your details

Forgot password? Click here to reset