Enhancing variational generation through self-decomposition

02/06/2022
by   Andrea Asperti, et al.
1

In this article we introduce the notion of Split Variational Autoencoder (SVAE), whose output x̂ is obtained as a weighted sum σ⊙x̂_̂1̂ + (1-σ) ⊙x̂_̂2̂ of two generated images x̂_̂1̂,x̂_̂2̂, and σ is a learned compositional map. The network is trained as a usual Variational Autoencoder with a negative loglikelihood loss between training and reconstructed images. The decomposition is nondeterministic, but follows two main schemes, that we may roughly categorize as either "syntactic" or "semantic". In the first case, the map tends to exploit the strong correlation between adjacent pixels, splitting the image in two complementary high frequency sub-images. In the second case, the map typically focuses on the contours of objects, splitting the image in interesting variations of its content, with more marked and distinctive features. In this case, the Fréchet Inception Distance (FID) of x̂_̂1̂ and x̂_̂2̂ is usually lower (hence better) than that of x̂, that clearly suffers from being the average of the formers. In a sense, a SVAE forces the Variational Autoencoder to make choices, in contrast with its intrinsic tendency to average between alternatives with the aim to minimize the reconstruction loss towards a specific sample. According to the FID metric, our technique, tested on typical datasets such as Mnist, Cifar10 and Celeba, allows us to outperform all previous purely variational architectures (not relying on normalization flows).

READ FULL TEXT

page 3

page 4

page 11

page 12

page 14

page 15

page 16

page 17

research
04/16/2020

Conditioned Variational Autoencoder for top-N item recommendation

In this paper, we propose a Conditioned Variational Autoencoder to impro...
research
07/11/2017

Least Square Variational Bayesian Autoencoder with Regularization

In recent years Variation Autoencoders have become one of the most popul...
research
11/22/2022

Using conditional variational autoencoders to generate images from atmospheric Cherenkov telescopes

High-energy particles hitting the upper atmosphere of the Earth produce ...
research
04/07/2021

Learning robust speech representation with an articulatory-regularized variational autoencoder

It is increasingly considered that human speech perception and productio...
research
05/14/2021

DoS and DDoS Mitigation Using Variational Autoencoders

DoS and DDoS attacks have been growing in size and number over the last ...
research
11/21/2020

Use of Student's t-Distribution for the Latent Layer in a Coupled Variational Autoencoder

A Coupled Variational Autoencoder, which incorporates both a generalized...
research
06/05/2018

Explaining Away Syntactic Structure in Semantic Document Representations

Most generative document models act on bag-of-words input in an attempt ...

Please sign up or login with your details

Forgot password? Click here to reset