RIGID: Recurrent GAN Inversion and Editing of Real Face Videos

08/11/2023
by   Yangyang Xu, et al.
0

GAN inversion is indispensable for applying the powerful editability of GAN to real images. However, existing methods invert video frames individually often leading to undesired inconsistent results over time. In this paper, we propose a unified recurrent framework, named Recurrent vIdeo GAN Inversion and eDiting (RIGID), to explicitly and simultaneously enforce temporally coherent GAN inversion and facial editing of real videos. Our approach models the temporal relations between current and previous frames from three aspects. To enable a faithful real video reconstruction, we first maximize the inversion fidelity and consistency by learning a temporal compensated latent code. Second, we observe incoherent noises lie in the high-frequency domain that can be disentangled from the latent space. Third, to remove the inconsistency after attribute manipulation, we propose an in-between frame composition constraint such that the arbitrary frame must be a direct composite of its neighboring frames. Our unified framework learns the inherent coherence between input frames in an end-to-end manner, and therefore it is agnostic to a specific attribute and can be applied to arbitrary editing of the same video without re-training. Extensive experiments demonstrate that RIGID outperforms state-of-the-art methods qualitatively and quantitatively in both inversion and editing tasks. The deliverables can be found in <https://cnnlstm.github.io/RIGID>

READ FULL TEXT

page 1

page 3

page 6

page 7

page 13

research
07/03/2020

Task-agnostic Temporally Consistent Facial Video Editing

Recent research has witnessed the advances in facial image editing tasks...
research
07/29/2021

From Continuity to Editability: Inverting GANs with Consecutive Images

Existing GAN inversion methods are stuck in a paradox that the inverted ...
research
10/18/2022

WaGI : Wavelet-based GAN Inversion for Preserving High-frequency Image Details

Recent GAN inversion models focus on preserving image-specific details t...
research
08/12/2021

UniFaceGAN: A Unified Framework for Temporally Consistent Facial Video Editing

Recent research has witnessed advances in facial image editing tasks inc...
research
12/19/2022

Photo-Realistic Out-of-domain GAN inversion via Invertibility Decomposition

The fidelity of Generative Adversarial Networks (GAN) inversion is imped...
research
02/09/2023

In-N-Out: Face Video Inversion and Editing with Volumetric Decomposition

3D-aware GANs offer new capabilities for creative content editing, such ...
research
03/28/2023

VIVE3D: Viewpoint-Independent Video Editing using 3D-Aware GANs

We introduce VIVE3D, a novel approach that extends the capabilities of i...

Please sign up or login with your details

Forgot password? Click here to reset