Learning Deep-Latent Hierarchies by Stacking Wasserstein Autoencoders

10/07/2020
by   Benoit Gaujac, et al.
0

Probabilistic models with hierarchical-latent-variable structures provide state-of-the-art results amongst non-autoregressive, unsupervised density-based models. However, the most common approach to training such models based on Variational Autoencoders (VAEs) often fails to leverage deep-latent hierarchies; successful approaches require complex inference and optimisation schemes. Optimal Transport is an alternative, non-likelihood-based framework for training generative models with appealing theoretical properties, in principle allowing easier training convergence between distributions. In this work we propose a novel approach to training models with deep-latent hierarchies based on Optimal Transport, without the need for highly bespoke models and inference networks. We show that our method enables the generative model to fully leverage its deep-latent hierarchy, avoiding the well known "latent variable collapse" issue of VAEs; therefore, providing qualitatively better sample generations as well as more interpretable latent representation than the original Wasserstein Autoencoder with Maximum Mean Discrepancy divergence.

READ FULL TEXT

page 5

page 7

page 8

page 12

research
10/04/2019

Stacked Wasserstein Autoencoder

Approximating distributions over complicated manifolds, such as natural ...
research
06/12/2018

Gaussian mixture models with Wasserstein distance

Generative models with both discrete and continuous latent variables are...
research
06/02/2022

Indeterminacy in Latent Variable Models: Characterization and Strong Identifiability

Most modern latent variable and probabilistic generative models, such as...
research
10/02/2018

Sinkhorn AutoEncoders

Optimal Transport offers an alternative to maximum likelihood for learni...
research
07/27/2019

Variational f-divergence Minimization

Probabilistic models are often trained by maximum likelihood, which corr...
research
07/09/2021

The Effects of Invertibility on the Representational Complexity of Encoders in Variational Autoencoders

Training and using modern neural-network based latent-variable generativ...
research
07/17/2022

Mean field Variational Inference via Wasserstein Gradient Flow

Variational inference (VI) provides an appealing alternative to traditio...

Please sign up or login with your details

Forgot password? Click here to reset