GANs N' Roses: Stable, Controllable, Diverse Image to Image Translation (works for videos too!)

06/11/2021
by   Min Jin Chong, et al.
41

We show how to learn a map that takes a content code, derived from a face image, and a randomly chosen style code to an anime image. We derive an adversarial loss from our simple and effective definitions of style and content. This adversarial loss guarantees the map is diverse – a very wide range of anime can be produced from a single content code. Under plausible assumptions, the map is not just diverse, but also correctly represents the probability of an anime, conditioned on an input face. In contrast, current multimodal generation procedures cannot capture the complex styles that appear in anime. Extensive quantitative experiments support the idea the map is correct. Extensive qualitative results show that the method can generate a much more diverse range of styles than SOTA comparisons. Finally, we show that our formalization of content and style allows us to perform video to video translation without ever training on videos.

READ FULL TEXT

page 2

page 4

page 6

page 7

research
03/02/2021

Image-to-image Translation via Hierarchical Style Disentanglement

Recently, image-to-image translation has made significant progress in ac...
research
12/01/2020

Unpaired Image-to-Image Translation via Latent Energy Transport

Image-to-image translation aims to preserve source contents while transl...
research
10/27/2021

Separating Content and Style for Unsupervised Image-to-Image Translation

Unsupervised image-to-image translation aims to learn the mapping betwee...
research
07/22/2022

Few-shot Image Generation Using Discrete Content Representation

Few-shot image generation and few-shot image translation are two related...
research
07/23/2021

Image-to-Image Translation with Low Resolution Conditioning

Most image-to-image translation methods focus on learning mappings acros...
research
11/21/2020

Stochastic Talking Face Generation Using Latent Distribution Matching

The ability to envisage the visual of a talking face based just on heari...
research
08/15/2018

Recycle-GAN: Unsupervised Video Retargeting

We introduce a data-driven approach for unsupervised video retargeting t...

Please sign up or login with your details

Forgot password? Click here to reset