StyleInV: A Temporal Style Modulated Inversion Network for Unconditional Video Generation

08/31/2023
by   Yuhan Wang, et al.
0

Unconditional video generation is a challenging task that involves synthesizing high-quality videos that are both coherent and of extended duration. To address this challenge, researchers have used pretrained StyleGAN image generators for high-quality frame synthesis and focused on motion generator design. The motion generator is trained in an autoregressive manner using heavy 3D convolutional discriminators to ensure motion coherence during video generation. In this paper, we introduce a novel motion generator design that uses a learning-based inversion network for GAN. The encoder in our method captures rich and smooth priors from encoding images to latents, and given the latent of an initially generated frame as guidance, our method can generate smooth future latent by modulating the inversion encoder temporally. Our method enjoys the advantage of sparse training and naturally constrains the generation space of our motion generator with the inversion network guided by the initial frame, eliminating the need for heavy discriminators. Moreover, our method supports style transfer with simple fine-tuning when the encoder is paired with a pretrained StyleGAN generator. Extensive experiments conducted on various benchmarks demonstrate the superiority of our method in generating long and high-resolution videos with decent single-frame quality and temporal consistency.

READ FULL TEXT

page 2

page 7

page 8

page 9

page 12

page 13

page 14

page 15

research
03/06/2023

MotionVideoGAN: A Novel Video Generator Based on the Motion Space Learned from Image Pairs

Video generation has achieved rapid progress benefiting from high-qualit...
research
12/11/2020

Intrinsic Temporal Regularization for High-resolution Human Video Synthesis

Temporal consistency is crucial for extending image processing pipelines...
research
04/30/2021

A Good Image Generator Is What You Need for High-Resolution Video Synthesis

Image and video synthesis are closely related areas aiming at generating...
research
10/17/2021

AE-StyleGAN: Improved Training of Style-Based Auto-Encoders

StyleGANs have shown impressive results on data generation and manipulat...
research
10/27/2022

Learning Variational Motion Prior for Video-based Motion Capture

Motion capture from a monocular video is fundamental and crucial for us ...
research
10/14/2021

Towards Using Clothes Style Transfer for Scenario-aware Person Video Generation

Clothes style transfer for person video generation is a challenging task...
research
10/10/2021

Sketch Me A Video

Video creation has been an attractive yet challenging task for artists t...

Please sign up or login with your details

Forgot password? Click here to reset