High Fidelity Image Synthesis With Deep VAEs In Latent Space

03/23/2023
by   Troy Luhman, et al.
0

We present fast, realistic image generation on high-resolution, multimodal datasets using hierarchical variational autoencoders (VAEs) trained on a deterministic autoencoder's latent space. In this two-stage setup, the autoencoder compresses the image into its semantic features, which are then modeled with a deep VAE. With this method, the VAE avoids modeling the fine-grained details that constitute the majority of the image's code length, allowing it to focus on learning its structural components. We demonstrate the effectiveness of our two-stage approach, achieving a FID of 9.34 on the ImageNet-256 dataset which is comparable to BigGAN. We make our implementation available online.

READ FULL TEXT

page 2

page 6

page 7

page 15

page 16

page 17

page 18

page 19

research
03/06/2023

Towards Composable Distributions of Latent Space Augmentations

We propose a composable framework for latent space image augmentation th...
research
06/02/2019

Generating Diverse High-Fidelity Images with VQ-VAE-2

We explore the use of Vector Quantized Variational AutoEncoder (VQ-VAE) ...
research
04/10/2023

Binary Latent Diffusion

In this paper, we show that a binary latent space can be explored for co...
research
03/01/2022

Variational Autoencoders Without the Variation

Variational autoencdoers (VAE) are a popular approach to generative mode...
research
03/25/2022

Efficient-VDVAE: Less is more

Hierarchical VAEs have emerged in recent years as a reliable option for ...
research
07/30/2021

Data-driven modeling of time-domain induced polarization

We present a novel approach for data-driven modeling of the time-domain ...
research
10/18/2022

Optimizing Hierarchical Image VAEs for Sample Quality

While hierarchical variational autoencoders (VAEs) have achieved great d...

Please sign up or login with your details

Forgot password? Click here to reset