An Improved Training Procedure for Neural Autoregressive Data Completion

11/23/2017
by   Maxime Voisin, et al.
0

Neural autoregressive models are explicit density estimators that achieve state-of-the-art likelihoods for generative modeling. The D-dimensional data distribution is factorized into an autoregressive product of one-dimensional conditional distributions according to the chain rule. Data completion is a more involved task than data generation: the model must infer missing variables for any partially observed input vector. Previous work introduced an order-agnostic training procedure for data completion with autoregressive models. Missing variables in any partially observed input vector can be imputed efficiently by choosing an ordering where observed dimensions precede unobserved ones and by computing the autoregressive product in this order. In this paper, we provide evidence that the order-agnostic (OA) training procedure is suboptimal for data completion. We propose an alternative procedure (OA++) that reaches better performance in fewer computations. It can handle all data completion queries while training fewer one-dimensional conditional distributions than the OA procedure. In addition, these one-dimensional conditional distributions are trained proportionally to their expected usage at inference time, reducing overfitting. Finally, our OA++ procedure can exploit prior knowledge about the distribution of inference completion queries, as opposed to OA. We support these claims with quantitative experiments on standard datasets used to evaluate autoregressive generative models.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/22/2020

Locally Masked Convolution for Autoregressive Models

High-dimensional generative models have many applications including imag...
research
06/13/2021

The DEformer: An Order-Agnostic Distribution Estimating Transformer

Order-agnostic autoregressive distribution estimation (OADE), i.e., auto...
research
01/16/2020

Masking schemes for universal marginalisers

We consider the effect of structure-agnostic and structure-dependent mas...
research
05/26/2022

Training and Inference on Any-Order Autoregressive Models the Right Way

Conditional inference on arbitrary subsets of variables is a core proble...
research
02/12/2015

MADE: Masked Autoencoder for Distribution Estimation

There has been a lot of recent interest in designing neural network mode...
research
12/01/2022

Sparsity Agnostic Depth Completion

We present a novel depth completion approach agnostic to the sparsity of...
research
03/22/2017

LogitBoost autoregressive networks

Multivariate binary distributions can be decomposed into products of uni...

Please sign up or login with your details

Forgot password? Click here to reset