NVAE: A Deep Hierarchical Variational Autoencoder

07/08/2020
by   Arash Vahdat, et al.
22

Normalizing flows, autoregressive models, variational autoencoders (VAEs), and deep energy-based models are among competing likelihood-based frameworks for deep generative learning. Among them, VAEs have the advantage of fast and tractable sampling and easy-to-access encoding networks. However, they are currently outperformed by other models such as normalizing flows and autoregressive models. While the majority of the research in VAEs is focused on the statistical challenges, we explore the orthogonal direction of carefully designing neural architectures for hierarchical VAEs. We propose Nouveau VAE (NVAE), a deep hierarchical VAE built for image generation using depth-wise separable convolutions and batch normalization. NVAE is equipped with a residual parameterization of Normal distributions and its training is stabilized by spectral regularization. We show that NVAE achieves state-of-the-art results among non-autoregressive likelihood-based models on the MNIST, CIFAR-10, and CelebA HQ datasets and it provides a strong baseline on FFHQ. For example, on CIFAR-10, NVAE pushes the state-of-the-art from 2.98 to 2.91 bits per dimension, and it produces high-quality images on CelebA HQ as shown in Fig. 1. To the best of our knowledge, NVAE is the first successful VAE applied to natural images as large as 256×256 pixels.

READ FULL TEXT

page 1

page 8

page 15

page 16

page 17

page 18

page 20

research
11/20/2020

Very Deep VAEs Generalize Autoregressive Models and Can Outperform Them on Images

We present a hierarchical VAE that, for the first time, outperforms the ...
research
07/20/2020

Incorporating Reinforced Adversarial Learning in Autoregressive Image Generation

Autoregressive models recently achieved comparable results versus state-...
research
04/05/2022

Split Hierarchical Variational Compression

Variational autoencoders (VAEs) have witnessed great success in performi...
research
01/30/2020

Learning Discrete Distributions by Dequantization

Media is generally stored digitally and is therefore discrete. Many succ...
research
10/01/2020

VAEBM: A Symbiosis between Variational Autoencoders and Energy-based Models

Energy-based models (EBMs) have recently been successful in representing...
research
03/01/2022

Variational Autoencoders Without the Variation

Variational autoencdoers (VAE) are a popular approach to generative mode...
research
03/05/2021

Generating Images with Sparse Representations

The high dimensionality of images presents architecture and sampling-eff...

Please sign up or login with your details

Forgot password? Click here to reset