AdaCat: Adaptive Categorical Discretization for Autoregressive Models

08/03/2022
by   Qiyang Li, et al.
4

Autoregressive generative models can estimate complex continuous data distributions, like trajectory rollouts in an RL environment, image intensities, and audio. Most state-of-the-art models discretize continuous data into several bins and use categorical distributions over the bins to approximate the continuous data distribution. The advantage is that the categorical distribution can easily express multiple modes and are straightforward to optimize. However, such approximation cannot express sharp changes in density without using significantly more bins, making it parameter inefficient. We propose an efficient, expressive, multimodal parameterization called Adaptive Categorical Discretization (AdaCat). AdaCat discretizes each dimension of an autoregressive model adaptively, which allows the model to allocate density to fine intervals of interest, improving parameter efficiency. AdaCat generalizes both categoricals and quantile-based regression. AdaCat is a simple add-on to any discretization-based distribution estimator. In experiments, AdaCat improves density estimation for real-world tabular data, images, audio, and trajectories, and improves planning in model-based offline RL.

READ FULL TEXT
research
06/11/2020

NanoFlow: Scalable Normalizing Flows with Sublinear Parameter Complexity

Normalizing flows (NFs) have become a prominent method for deep generati...
research
03/08/2023

Diffusing Gaussian Mixtures for Generating Categorical Data

Learning a categorical distribution comes with its own set of challenges...
research
02/10/2021

Argmax Flows and Multinomial Diffusion: Towards Non-Autoregressive Language Models

The field of language modelling has been largely dominated by autoregres...
research
06/17/2020

Categorical Normalizing Flows via Continuous Transformations

Despite their popularity, to date, the application of normalizing flows ...
research
10/27/2017

Few-shot Autoregressive Density Estimation: Towards Learning to Learn Distributions

Deep autoregressive models have shown state-of-the-art performance in de...
research
01/30/2020

Learning Discrete Distributions by Dequantization

Media is generally stored digitally and is therefore discrete. Many succ...
research
04/12/2017

A Neural Parametric Singing Synthesizer

We present a new model for singing synthesis based on a modified version...

Please sign up or login with your details

Forgot password? Click here to reset