Improving Diversity in Zero-Shot GAN Adaptation with Semantic Variations

08/21/2023
by   Seogkyu Jeon, et al.
0

Training deep generative models usually requires a large amount of data. To alleviate the data collection cost, the task of zero-shot GAN adaptation aims to reuse well-trained generators to synthesize images of an unseen target domain without any further training samples. Due to the data absence, the textual description of the target domain and the vision-language models, e.g., CLIP, are utilized to effectively guide the generator. However, with only a single representative text feature instead of real images, the synthesized images gradually lose diversity as the model is optimized, which is also known as mode collapse. To tackle the problem, we propose a novel method to find semantic variations of the target text in the CLIP space. Specifically, we explore diverse semantic variations based on the informative text feature of the target domain while regularizing the uncontrolled deviation of the semantic information. With the obtained variations, we design a novel directional moment loss that matches the first and second moments of image and text direction distributions. Moreover, we introduce elastic weight consolidation and a relation consistency loss to effectively preserve valuable content information from the source domain, e.g., appearances. Through extensive experiments, we demonstrate the efficacy of the proposed methods in ensuring sample diversity in various scenarios of zero-shot GAN adaptation. We also conduct ablation studies to validate the effect of each proposed component. Notably, our model achieves a new state-of-the-art on zero-shot GAN adaptation in terms of both diversity and quality.

READ FULL TEXT

page 2

page 3

page 6

page 7

page 8

page 12

page 18

page 20

research
04/06/2023

Zero-shot Generative Model Adaptation via Image-specific Prompt Learning

Recently, CLIP-guided image synthesis has shown appealing performance on...
research
02/25/2021

Domain Adaptation for Learning Generator from Paired Few-Shot Data

We propose a Paired Few-shot GAN (PFS-GAN) model for learning generators...
research
03/19/2021

Paint by Word

We investigate the problem of zero-shot semantic image painting. Instead...
research
11/29/2022

DATID-3D: Diversity-Preserved Domain Adaptation Using Text-to-Image Diffusion for 3D Generative Model

Recent 3D generative models have achieved remarkable performance in synt...
research
11/23/2016

Learning Joint Feature Adaptation for Zero-Shot Recognition

Zero-shot recognition (ZSR) aims to recognize target-domain data instanc...
research
05/08/2022

A Closer Look at Few-shot Image Generation

Modern GANs excel at generating high quality and diverse images. However...
research
11/25/2022

Expanding Small-Scale Datasets with Guided Imagination

The power of Deep Neural Networks (DNNs) depends heavily on the training...

Please sign up or login with your details

Forgot password? Click here to reset