Unsupervised multi-modal Styled Content Generation

01/10/2020
by   Omry Sendik, et al.
0

The emergence of deep generative models has recently enabled the automatic generation of massive amounts of graphical content, both in 2D and in 3D. Generative Adversarial Networks (GANs) and style control mechanisms, such as Adaptive Instance Normalization (AdaIN), have proved particularly effective in this context, culminating in the state-of-the-art StyleGAN architecture. While such models are able to learn diverse distributions, provided a sufficiently large training set, they are not well-suited for scenarios where the distribution of the training data exhibits a multi-modal behavior. In such cases, reshaping a uniform or normal distribution over the latent space into a complex multi-modal distribution in the data domain is challenging, and the generator might fail to sample the target distribution well. Furthermore, existing unsupervised generative models are not able to control the mode of the generated samples independently of the other visual attributes, despite the fact that they are typically disentangled in the training data. In this paper, we introduce UMMGAN, a novel architecture designed to better model multi-modal distributions, in an unsupervised fashion. Building upon the StyleGAN architecture, our network learns multiple modes, in a completely unsupervised manner, and combines them using a set of learned weights. We demonstrate that this approach is capable of effectively approximating a complex distribution as a superposition of multiple simple ones. We further show that UMMGAN effectively disentangles between modes and style, thereby providing an independent degree of control over the generated content.

READ FULL TEXT
research
01/10/2020

Unsupervised K-modal Styled Content Generation

The emergence of generative models based on deep neural networks has rec...
research
11/15/2019

MMGAN: Generative Adversarial Networks for Multi-Modal Distributions

Over the past years, Generative Adversarial Networks (GANs) have shown a...
research
08/11/2019

GAN-Tree: An Incrementally Learned Hierarchical Generative Framework for Multi-Modal Data Distributions

Despite the remarkable success of generative adversarial networks, their...
research
01/25/2019

Diversity-Sensitive Conditional Generative Adversarial Networks

We propose a simple yet highly effective method that addresses the mode-...
research
04/14/2021

StEP: Style-based Encoder Pre-training for Multi-modal Image Synthesis

We propose a novel approach for multi-modal Image-to-image (I2I) transla...
research
11/28/2018

Play as You Like: Timbre-enhanced Multi-modal Music Style Transfer

Style transfer of polyphonic music recordings is a challenging task when...
research
04/01/2020

A theory of independent mechanisms for extrapolation in generative models

Deep generative models reproduce complex empirical data but cannot extra...

Please sign up or login with your details

Forgot password? Click here to reset