Re-parameterizing VAEs for stability

06/25/2021
by   David Dehaene, et al.
0

We propose a theoretical approach towards the training numerical stability of Variational AutoEncoders (VAE). Our work is motivated by recent studies empowering VAEs to reach state of the art generative results on complex image datasets. These very deep VAE architectures, as well as VAEs using more complex output distributions, highlight a tendency to haphazardly produce high training gradients as well as NaN losses. The empirical fixes proposed to train them despite their limitations are neither fully theoretically grounded nor generally sufficient in practice. Building on this, we localize the source of the problem at the interface between the model's neural networks and their output probabilistic distributions. We explain a common source of instability stemming from an incautious formulation of the encoded Normal distribution's variance, and apply the same approach on other, less obvious sources. We show that by implementing small changes to the way we parameterize the Normal distributions on which they rely, VAEs can securely be trained.

READ FULL TEXT
research
02/23/2020

Variance Loss in Variational Autoencoders

In this article, we highlight what appears to be major issue of Variatio...
research
10/16/2019

Optimal Transport Based Generative Autoencoders

The field of deep generative modeling is dominated by generative adversa...
research
05/14/2021

Adapting deep generative approaches for getting synthetic data with realistic marginal distributions

Synthetic data generation is of great interest in diverse applications, ...
research
11/27/2020

Direct Evolutionary Optimization of Variational Autoencoders With Binary Latents

Discrete latent variables are considered important for real world data, ...
research
02/12/2021

Demystifying Inductive Biases for β-VAE Based Architectures

The performance of β-Variational-Autoencoders (β-VAEs) and their variant...
research
07/20/2021

ByPE-VAE: Bayesian Pseudocoresets Exemplar VAE

Recent studies show that advanced priors play a major role in deep gener...
research
10/05/2021

Top-N: Equivariant set and graph generation without exchangeability

We consider one-shot probabilistic decoders that map a vector-shaped pri...

Please sign up or login with your details

Forgot password? Click here to reset