Zero-shot Generative Model Adaptation via Image-specific Prompt Learning

04/06/2023
by   Jiayi Guo, et al.
0

Recently, CLIP-guided image synthesis has shown appealing performance on adapting a pre-trained source-domain generator to an unseen target domain. It does not require any target-domain samples but only the textual domain labels. The training is highly efficient, e.g., a few minutes. However, existing methods still have some limitations in the quality of generated images and may suffer from the mode collapse issue. A key reason is that a fixed adaptation direction is applied for all cross-domain image pairs, which leads to identical supervision signals. To address this issue, we propose an Image-specific Prompt Learning (IPL) method, which learns specific prompt vectors for each source-domain image. This produces a more precise adaptation direction for every cross-domain image pair, endowing the target-domain generator with greatly enhanced flexibility. Qualitative and quantitative evaluations on various domains demonstrate that IPL effectively improves the quality and diversity of synthesized images and alleviates the mode collapse. Moreover, IPL is independent of the structure of the generative model, such as generative adversarial networks or diffusion models. Code is available at https://github.com/Picsart-AI-Research/IPL-Zero-Shot-Generative-Model-Adaptation.

READ FULL TEXT

page 1

page 6

page 7

page 8

page 12

page 13

page 14

page 15

research
07/18/2022

Towards Diverse and Faithful One-shot Adaption of Generative Adversarial Networks

One-shot generative domain adaption aims to transfer a pre-trained gener...
research
02/25/2021

Domain Adaptation for Learning Generator from Paired Few-Shot Data

We propose a Paired Few-shot GAN (PFS-GAN) model for learning generators...
research
06/03/2018

NAM: Non-Adversarial Unsupervised Domain Mapping

Several methods were recently proposed for the task of translating image...
research
08/21/2023

Improving Diversity in Zero-Shot GAN Adaptation with Semantic Variations

Training deep generative models usually requires a large amount of data....
research
08/01/2023

Domain Adaptation based on Human Feedback for Enhancing Generative Model Denoising Abilities

How can we apply human feedback into generative model? As answer of this...
research
01/31/2023

Zero-shot-Learning Cross-Modality Data Translation Through Mutual Information Guided Stochastic Diffusion

Cross-modality data translation has attracted great interest in image co...
research
02/06/2023

Domain Re-Modulation for Few-Shot Generative Domain Adaptation

In this study, we investigate the task of few-shot Generative Domain Ada...

Please sign up or login with your details

Forgot password? Click here to reset