Efficient-VDVAE: Less is more

03/25/2022
by   Louay Hazami, et al.
6

Hierarchical VAEs have emerged in recent years as a reliable option for maximum likelihood estimation. However, instability issues and demanding computational requirements have hindered research progress in the area. We present simple modifications to the Very Deep VAE to make it converge up to 2.6× faster, save up to 20× in memory load and improve stability during training. Despite these changes, our models achieve comparable or better negative log-likelihood performance than current state-of-the-art models on all 7 commonly used image datasets we evaluated on. We also make an argument against using 5-bit benchmarks as a way to measure hierarchical VAE's performance due to undesirable biases caused by the 5-bit quantization. Additionally, we empirically demonstrate that roughly 3% of the hierarchical VAE's latent space dimensions is sufficient to encode most of the image information, without loss of performance, opening up the doors to efficiently leverage the hierarchical VAEs' latent space in downstream tasks. We release our source code and models at https://github.com/Rayhane-mamah/Efficient-VDVAE .

READ FULL TEXT

page 6

page 17

page 20

page 21

page 22

page 23

page 24

page 25

research
11/20/2020

Very Deep VAEs Generalize Autoregressive Models and Can Outperform Them on Images

We present a hierarchical VAE that, for the first time, outperforms the ...
research
10/18/2022

Optimizing Hierarchical Image VAEs for Sample Quality

While hierarchical variational autoencoders (VAEs) have achieved great d...
research
03/23/2023

High Fidelity Image Synthesis With Deep VAEs In Latent Space

We present fast, realistic image generation on high-resolution, multimod...
research
10/05/2020

Unbiased Gradient Estimation for Variational Auto-Encoders using Coupled Markov Chains

The variational auto-encoder (VAE) is a deep latent variable model that ...
research
07/03/2020

Variational Autoencoders for Anomalous Jet Tagging

We present a detailed study on Variational Autoencoders (VAEs) for anoma...
research
01/19/2017

PixelCNN++: Improving the PixelCNN with Discretized Logistic Mixture Likelihood and Other Modifications

PixelCNNs are a recently proposed class of powerful generative models wi...
research
03/16/2021

Spatial Dependency Networks: Neural Layers for Improved Generative Image Modeling

How to improve generative modeling by better exploiting spatial regulari...

Please sign up or login with your details

Forgot password? Click here to reset