A Batch Normalized Inference Network Keeps the KL Vanishing Away

04/27/2020
by   Qile Zhu, et al.
0

Variational Autoencoder (VAE) is widely used as a generative model to approximate a model's posterior on latent variables by combining the amortized variational inference and deep neural networks. However, when paired with strong autoregressive decoders, VAE often converges to a degenerated local optimum known as "posterior collapse". Previous approaches consider the Kullback Leibler divergence (KL) individual for each datapoint. We propose to let the KL follow a distribution across the whole dataset, and analyze that it is sufficient to prevent posterior collapse by keeping the expectation of the KL's distribution positive. Then we propose Batch Normalized-VAE (BN-VAE), a simple but effective approach to set a lower bound of the expectation by regularizing the distribution of the approximate posterior's parameters. Without introducing any new model component or modifying the objective, our approach can avoid the posterior collapse effectively and efficiently. We further show that the proposed BN-VAE can be extended to conditional VAE (CVAE). Empirically, our approach surpasses strong autoregressive baselines on language modeling, text classification and dialogue generation, and rivals more complex approaches while keeping almost the same training time as VAE.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
11/02/2020

Improving Variational Autoencoder for Text Modelling with Timestep-Wise Regularisation

The Variational Autoencoder (VAE) is a popular and powerful model applie...
research
09/13/2019

ρ-VAE: Autoregressive parametrization of the VAE encoder

We make a minimal, but very effective alteration to the VAE model. This ...
research
10/31/2020

ControlVAE: Tuning, Analytical Properties, and Performance Analysis

This paper reviews the novel concept of controllable variational autoenc...
research
06/29/2020

VAE-KRnet and its applications to variational Bayes

In this work, we have proposed a generative model for density estimation...
research
12/23/2019

The Usual Suspects? Reassessing Blame for VAE Posterior Collapse

In narrow asymptotic settings Gaussian VAE models of continuous data hav...
research
03/11/2022

Dual reparametrized Variational Generative Model for Time-Series Forecasting

This paper propose DualVDT, a generative model for Time-series forecasti...
research
09/02/2019

A Surprisingly Effective Fix for Deep Latent Variable Modeling of Text

When trained effectively, the Variational Autoencoder (VAE) is both a po...

Please sign up or login with your details

Forgot password? Click here to reset