Learning Discrete Distributions by Dequantization

01/30/2020
by   Emiel Hoogeboom, et al.
16

Media is generally stored digitally and is therefore discrete. Many successful deep distribution models in deep learning learn a density, i.e., the distribution of a continuous random variable. Naïve optimization on discrete data leads to arbitrarily high likelihoods, and instead, it has become standard practice to add noise to datapoints. In this paper, we present a general framework for dequantization that captures existing methods as a special case. We derive two new dequantization objectives: importance-weighted (iw) dequantization and Rényi dequantization. In addition, we introduce autoregressive dequantization (ARD) for more flexible dequantization distributions. Empirically we find that iw and Rényi dequantization considerably improve performance for uniform dequantization distributions. ARD achieves a negative log-likelihood of 3.06 bits per dimension on CIFAR10, which to the best of our knowledge is state-of-the-art among distribution models that do not require autoregressive inverses for sampling.

READ FULL TEXT

page 2

page 6

page 11

page 12

page 13

research
05/24/2019

Discrete Flows: Invertible Generative Models of Discrete Data

While normalizing flows have led to significant advances in modeling hig...
research
05/17/2021

Parallel and Flexible Sampling from Autoregressive Models via Langevin Dynamics

This paper introduces an alternative approach to sampling from autoregre...
research
07/08/2020

NVAE: A Deep Hierarchical Variational Autoencoder

Normalizing flows, autoregressive models, variational autoencoders (VAEs...
research
04/11/2019

Autoregressive Energy Machines

Neural density estimators are flexible families of parametric models whi...
research
08/03/2022

AdaCat: Adaptive Categorical Discretization for Autoregressive Models

Autoregressive generative models can estimate complex continuous data di...
research
03/18/2016

Document Neural Autoregressive Distribution Estimation

We present an approach based on feed-forward neural networks for learnin...
research
04/03/2019

Creating new distributions using integration and summation by parts

Methods for generating new distributions from old can be thought of as t...

Please sign up or login with your details

Forgot password? Click here to reset