Log In Sign Up

Swapping Autoencoder for Deep Image Manipulation

by   Taesung Park, et al.

Deep generative models have become increasingly effective at producing realistic images from randomly sampled seeds, but using such models for controllable manipulation of existing images remains challenging. We propose the Swapping Autoencoder, a deep model designed specifically for image manipulation, rather than random sampling. The key idea is to encode an image with two independent components and enforce that any swapped combination maps to a realistic image. In particular, we encourage the components to represent structure and texture, by enforcing one component to encode co-occurrent patch statistics across different parts of an image. As our method is trained with an encoder, finding the latent codes for a new input image becomes trivial, rather than cumbersome. As a result, it can be used to manipulate real input images in various ways, including texture swapping, local and global editing, and latent code vector arithmetic. Experiments on multiple datasets show that our model produces better results and is substantially more efficient compared to recent generative models.


page 5

page 8

page 9

page 16

page 17

page 18

page 19

page 20


Text and Image Guided 3D Avatar Generation and Manipulation

The manipulation of latent space has recently become an interesting topi...

EdiBERT, a generative model for image editing

Advances in computer vision are pushing the limits of im-age manipulatio...

Proactive Image Manipulation Detection

Image manipulation detection algorithms are often trained to discriminat...

Towards Counterfactual Image Manipulation via CLIP

Leveraging StyleGAN's expressivity and its disentangled latent codes, ex...

Controllable cardiac synthesis via disentangled anatomy arithmetic

Acquiring annotated data at scale with rare diseases or conditions remai...

Learning to Manipulate Individual Objects in an Image

We describe a method to train a generative model with latent factors tha...