Image Disentanglement and Uncooperative Re-Entanglement for High-Fidelity Image-to-Image Translation

01/11/2019
by   Adam W. Harley, et al.
0

Cross-domain image-to-image translation should satisfy two requirements: (1) preserve the information that is common to both domains, and (2) generate convincing images covering variations that appear in the target domain. This is challenging, especially when there are no example translations available as supervision. Adversarial cycle consistency was recently proposed as a solution, with beautiful and creative results, yielding much follow-up work. However, augmented reality applications cannot readily use such techniques to provide users with compelling translations of real scenes, because the translations do not have high-fidelity constraints. In other words, current models are liable to change details that should be preserved: while re-texturing a face, they may alter the face's expression in an unpredictable way. In this paper, we introduce the problem of high-fidelity image-to-image translation, and present a method for solving it. Our main insight is that low-fidelity translations typically escape a cycle-consistency penalty, because the back-translator learns to compensate for the forward-translator's errors. We therefore introduce an optimization technique that prevents the networks from cooperating: simply train each network only when its input data is real. Prior works, in comparison, train each network with a mix of real and generated data. Experimental results show that our method accurately disentangles the factors that separate the domains, and converges to semantics-preserving translations that prior methods miss.

READ FULL TEXT

page 1

page 6

page 8

page 9

page 10

page 12

research
07/23/2019

Controlling biases and diversity in diverse image-to-image translation

The task of unpaired image-to-image translation is highly challenging du...
research
07/06/2020

MCMI: Multi-Cycle Image Translation with Mutual Information Constraints

We present a mutual information-based framework for unsupervised image-t...
research
09/26/2021

ISF-GAN: An Implicit Style Function for High-Resolution Image-to-Image Translation

Recently, there has been an increasing interest in image editing methods...
research
09/09/2021

Leveraging Local Domains for Image-to-Image Translation

Image-to-image (i2i) networks struggle to capture local changes because ...
research
03/08/2019

Mix and match networks: multi-domain alignment for unpaired image-to-image translation

This paper addresses the problem of inferring unseen cross-domain and cr...
research
12/03/2021

Image-to-image Translation as a Unique Source of Knowledge

Image-to-image (I2I) translation is an established way of translating da...
research
06/07/2023

GP-UNIT: Generative Prior for Versatile Unsupervised Image-to-Image Translation

Recent advances in deep learning have witnessed many successful unsuperv...

Please sign up or login with your details

Forgot password? Click here to reset