ALADIN-NST: Self-supervised disentangled representation learning of artistic style through Neural Style Transfer

04/12/2023
by   Dan Ruta, et al.
0

Representation learning aims to discover individual salient features of a domain in a compact and descriptive form that strongly identifies the unique characteristics of a given sample respective to its domain. Existing works in visual style representation literature have tried to disentangle style from content during training explicitly. A complete separation between these has yet to be fully achieved. Our paper aims to learn a representation of visual artistic style more strongly disentangled from the semantic content depicted in an image. We use Neural Style Transfer (NST) to measure and drive the learning signal and achieve state-of-the-art representation learning on explicitly disentangled metrics. We show that strongly addressing the disentanglement of style and content leads to large gains in style-specific metrics, encoding far less semantic information and achieving state-of-the-art accuracy in downstream multimodal applications.

READ FULL TEXT

page 2

page 3

page 6

page 7

research
08/13/2018

Disentangled Representation Learning for Text Style Transfer

This paper tackles the problem of disentangling the latent variables of ...
research
08/15/2023

StyleDiffusion: Controllable Disentangled Style Transfer via Diffusion Models

Content and style (C-S) disentanglement is a fundamental problem and cri...
research
08/05/2020

Domain-Specific Mappings for Generative Adversarial Style Transfer

Style transfer generates an image whose content comes from one image and...
research
05/11/2019

Disentangling Content and Style via Unsupervised Geometry Distillation

It is challenging to disentangle an object into two orthogonal spaces of...
research
06/01/2020

Improving Disentangled Text Representation Learning with Information-Theoretic Guidance

Learning disentangled representations of natural language is essential f...
research
12/16/2020

Multi-type Disentanglement without Adversarial Training

Controlling the style of natural language by disentangling the latent sp...
research
07/10/2023

Substance or Style: What Does Your Image Embedding Know?

Probes are small networks that predict properties of underlying data fro...

Please sign up or login with your details

Forgot password? Click here to reset