StyleDiffusion: Controllable Disentangled Style Transfer via Diffusion Models

08/15/2023
by   Zhizhong Wang, et al.
0

Content and style (C-S) disentanglement is a fundamental problem and critical challenge of style transfer. Existing approaches based on explicit definitions (e.g., Gram matrix) or implicit learning (e.g., GANs) are neither interpretable nor easy to control, resulting in entangled representations and less satisfying results. In this paper, we propose a new C-S disentangled framework for style transfer without using previous assumptions. The key insight is to explicitly extract the content information and implicitly learn the complementary style information, yielding interpretable and controllable C-S disentanglement and style transfer. A simple yet effective CLIP-based style disentanglement loss coordinated with a style reconstruction prior is introduced to disentangle C-S in the CLIP image space. By further leveraging the powerful style removal and generative ability of diffusion models, our framework achieves superior results than state of the art and flexible C-S disentanglement and trade-off control. Our work provides new insights into the C-S disentanglement in style transfer and demonstrates the potential of diffusion models for learning well-disentangled C-S characteristics.

READ FULL TEXT

page 6

page 9

page 18

page 19

page 21

page 22

page 23

page 24

research
06/07/2023

Interpretable Style Transfer for Text-to-Speech with ControlVAE and Diffusion Bridge

With the demand for autonomous control and personalized speech generatio...
research
06/15/2023

ArtFusion: Controllable Arbitrary Style Transfer using Dual Conditional Latent Diffusion Models

Arbitrary Style Transfer (AST) aims to transform images by adopting the ...
research
08/24/2020

CA-GAN: Weakly Supervised Color Aware GAN for Controllable Makeup Transfer

While existing makeup style transfer models perform an image synthesis w...
research
04/12/2023

ALADIN-NST: Self-supervised disentangled representation learning of artistic style through Neural Style Transfer

Representation learning aims to discover individual salient features of ...
research
10/14/2022

Controllable Style Transfer via Test-time Training of Implicit Neural Representation

We propose a controllable style transfer framework based on Implicit Neu...
research
05/25/2023

DDDM-VC: Decoupled Denoising Diffusion Models with Disentangled Representation and Prior Mixup for Verified Robust Voice Conversion

Diffusion-based generative models have exhibited powerful generative per...
research
02/20/2018

Stroke Controllable Fast Style Transfer with Adaptive Receptive Fields

Recently, in the community of Neural Style Transfer, several algorithms ...

Please sign up or login with your details

Forgot password? Click here to reset