Invertible Gaussian Reparameterization: Revisiting the Gumbel-Softmax

12/19/2019
by   Andres Potapczynski, et al.
36

The Gumbel-Softmax is a continuous distribution over the simplex that is often used as a relaxation of discrete distributions. Because it can be readily interpreted and easily reparameterized, the Gumbel-Softmax enjoys widespread use. We show that this relaxation experiences two shortcomings that affect its performance, namely: numerical instability caused by its temperature hyperparameter and noisy KL estimates. The first requires the temperature values to be set too high, creating a poor correspondence between continuous components and their respective discrete complements. The second, which is of fundamental importance to variational autoencoders, severely hurts performance. We propose a flexible and reparameterizable family of distributions that circumvents these issues by transforming Gaussian noise into one-hot approximations through an invertible function. Our construction improves numerical stability, and outperforms the Gumbel-Softmax in a variety of experiments while generating samples that are closer to their discrete counterparts and achieving lower-variance gradients. Furthermore, with a careful choice of the invertible function we extend the reparameterization trick to distributions with countably infinite support.

READ FULL TEXT

page 12

page 13

research
11/12/2016

GANS for Sequences of Discrete Elements with the Gumbel-softmax Distribution

Generative Adversarial Networks (GAN) have limitations when the goal is ...
research
03/04/2020

Generalized Gumbel-Softmax Gradient Estimator for Various Discrete Random Variables

Estimating the gradients of stochastic nodes is one of the crucial resea...
research
09/29/2018

Improved Gradient-Based Optimization Over Discrete Distributions

In many applications we seek to maximize an expectation with respect to ...
research
07/26/2023

Efficient Learning of Discrete-Continuous Computation Graphs

Numerous models for supervised and reinforcement learning benefit from c...
research
10/14/2020

Temperature check: theory and practice for training models with softmax-cross-entropy losses

The softmax function combined with a cross-entropy loss is a principled ...
research
08/04/2021

Sparse Continuous Distributions and Fenchel-Young Losses

Exponential families are widely used in machine learning; they include m...
research
12/22/2017

A short variational proof of equivalence between policy gradients and soft Q learning

Two main families of reinforcement learning algorithms, Q-learning and p...

Please sign up or login with your details

Forgot password? Click here to reset