Control4D: Dynamic Portrait Editing by Learning 4D GAN from 2D Diffusion-based Editor

05/31/2023
by   Ruizhi Shao, et al.
0

Recent years have witnessed considerable achievements in editing images with text instructions. When applying these editors to dynamic scene editing, the new-style scene tends to be temporally inconsistent due to the frame-by-frame nature of these 2D editors. To tackle this issue, we propose Control4D, a novel approach for high-fidelity and temporally consistent 4D portrait editing. Control4D is built upon an efficient 4D representation with a 2D diffusion-based editor. Instead of using direct supervisions from the editor, our method learns a 4D GAN from it and avoids the inconsistent supervision signals. Specifically, we employ a discriminator to learn the generation distribution based on the edited images and then update the generator with the discrimination signals. For more stable training, multi-level information is extracted from the edited images and used to facilitate the learning of the generator. Experimental results show that Control4D surpasses previous approaches and achieves more photo-realistic and consistent 4D editing performances. The link to our project website is https://control4darxiv.github.io.

READ FULL TEXT

page 1

page 4

page 6

page 7

page 8

research
06/21/2022

Temporally Consistent Semantic Video Editing

Generative adversarial networks (GANs) have demonstrated impressive imag...
research
08/28/2023

MagicEdit: High-Fidelity and Temporally Coherent Video Editing

In this report, we present MagicEdit, a surprisingly simple yet effectiv...
research
05/18/2023

DiffUTE: Universal Text Editing Diffusion Model

Diffusion model based language-guided image editing has achieved great s...
research
07/03/2020

Task-agnostic Temporally Consistent Facial Video Editing

Recent research has witnessed the advances in facial image editing tasks...
research
08/12/2021

UniFaceGAN: A Unified Framework for Temporally Consistent Facial Video Editing

Recent research has witnessed advances in facial image editing tasks inc...
research
03/28/2023

VIVE3D: Viewpoint-Independent Video Editing using 3D-Aware GANs

We introduce VIVE3D, a novel approach that extends the capabilities of i...
research
04/30/2022

Look Closer to Supervise Better: One-Shot Font Generation via Component-Based Discriminator

Automatic font generation remains a challenging research issue due to th...

Please sign up or login with your details

Forgot password? Click here to reset