Learning Input-agnostic Manipulation Directions in StyleGAN with Text Guidance

02/26/2023
by   Yoonjeon Kim, et al.
0

With the advantages of fast inference and human-friendly flexible manipulation, image-agnostic style manipulation via text guidance enables new applications that were not previously available. The state-of-the-art text-guided image-agnostic manipulation method embeds the representation of each channel of StyleGAN independently in the Contrastive Language-Image Pre-training (CLIP) space, and provides it in the form of a Dictionary to quickly find out the channel-wise manipulation direction during inference time. However, in this paper we argue that this dictionary which is constructed by controlling single channel individually is limited to accommodate the versatility of text guidance since the collective and interactive relation among multiple channels are not considered. Indeed, we show that it fails to discover a large portion of manipulation directions that can be found by existing methods, which manually manipulates latent space without texts. To alleviate this issue, we propose a novel method that learns a Dictionary, whose entry corresponds to the representation of a single channel, by taking into account the manipulation effect coming from the interaction with multiple other channels. We demonstrate that our strategy resolves the inability of previous methods in finding diverse known directions from unsupervised methods and unknown directions from random text while maintaining the real-time inference speed and disentanglement ability.

READ FULL TEXT

page 14

page 17

page 18

page 19

page 20

page 21

page 22

page 23

research
03/31/2021

StyleCLIP: Text-Driven Manipulation of StyleGAN Imagery

Inspired by the ability of StyleGAN to generate highly realistic images ...
research
12/05/2022

CLIPVG: Text-Guided Image Manipulation Using Differentiable Vector Graphics

Considerable progress has recently been made in leveraging CLIP (Contras...
research
12/15/2021

StyleMC: Multi-Channel Based Fast Text-Guided Image Generation and Manipulation

Discovering meaningful directions in the latent space of GANs to manipul...
research
10/10/2022

Bridging CLIP and StyleGAN through Latent Alignment for Image Editing

Text-driven image manipulation is developed since the vision-language mo...
research
02/18/2023

Attribute-Specific Manipulation Based on Layer-Wise Channels

Image manipulation on the latent space of the pre-trained StyleGAN can c...
research
05/26/2023

StyleHumanCLIP: Text-guided Garment Manipulation for StyleGAN-Human

This paper tackles text-guided control of StyleGAN for editing garments ...
research
12/20/2020

Latent Compass: Creation by Navigation

In Marius von Senden's Space and Sight, a newly sighted blind patient de...

Please sign up or login with your details

Forgot password? Click here to reset