TR0N: Translator Networks for 0-Shot Plug-and-Play Conditional Generation

04/26/2023
by   Zhaoyan Liu, et al.
0

We propose TR0N, a highly general framework to turn pre-trained unconditional generative models, such as GANs and VAEs, into conditional models. The conditioning can be highly arbitrary, and requires only a pre-trained auxiliary model. For example, we show how to turn unconditional models into class-conditional ones with the help of a classifier, and also into text-to-image models by leveraging CLIP. TR0N learns a lightweight stochastic mapping which "translates" between the space of conditions and the latent space of the generative model, in such a way that the generated latent corresponds to a data sample satisfying the desired condition. The translated latent samples are then further improved upon through Langevin dynamics, enabling us to obtain higher-quality data samples. TR0N requires no training data nor fine-tuning, yet can achieve a zero-shot FID of 10.9 on MS-COCO, outperforming competing alternatives not only on this metric, but also in sampling speed – all while retaining a much higher level of generality. Our code is available at https://github.com/layer6ai-labs/tr0n.

READ FULL TEXT

page 1

page 6

page 7

page 8

page 18

page 19

page 20

page 21

research
06/15/2020

Exponential Tilting of Generative Models: Improving Sample Quality by Training and Sampling from Latent Energy

In this paper, we present a general method that can improve the sample q...
research
08/31/2023

Ref-Diff: Zero-shot Referring Image Segmentation with Generative Models

Zero-shot referring image segmentation is a challenging task because it ...
research
09/14/2022

Generative Visual Prompt: Unifying Distributional Control of Pre-Trained Generative Models

Generative models (e.g., GANs and diffusion models) learn the underlying...
research
08/21/2023

Image-free Classifier Injection for Zero-Shot Classification

Zero-shot learning models achieve remarkable results on image classifica...
research
12/02/2021

FuseDream: Training-Free Text-to-Image Generation with Improved CLIP+GAN Space Optimization

Generating images from natural language instructions is an intriguing ye...
research
11/15/2017

Latent Constraints: Learning to Generate Conditionally from Unconditional Generative Models

Deep generative neural networks have proven effective at both conditiona...
research
06/29/2022

Forgetting Data from Pre-trained GANs

Large pre-trained generative models are known to occasionally provide sa...

Please sign up or login with your details

Forgot password? Click here to reset