Depthwise Discrete Representation Learning

04/11/2020
by   Iordanis Fostiropoulos, et al.
5

Recent advancements in learning Discrete Representations as opposed to continuous ones have led to state of art results in tasks that involve Language, Audio and Vision. Some latent factors such as words, phonemes and shapes are better represented by discrete latent variables as opposed to continuous. Vector Quantized Variational Autoencoders (VQVAE) have produced remarkable results in multiple domains. VQVAE learns a prior distribution z_e along with its mapping to a discrete number of K vectors (Vector Quantization). We propose applying VQ along the feature axis. We hypothesize that by doing so, we are learning a mapping between the codebook vectors and the marginal distribution of the prior feature space. Our approach leads to 33% improvement as compared to prevous discrete models and has similar performance to state of the art auto-regressive models (e.g. PixelSNAIL). We evaluate our approach on a static prior using an artificial toy dataset (blobs). We further evaluate our approach on benchmarks for CIFAR-10 and ImageNet.

READ FULL TEXT

page 12

page 17

page 18

page 19

page 20

research
09/07/2016

Discrete Variational Autoencoders

Probabilistic models with discrete latent variables naturally capture da...
research
06/13/2023

Vector-Quantized Graph Auto-Encoder

In this work, we addresses the problem of modeling distributions of grap...
research
05/18/2020

Robust Training of Vector Quantized Bottleneck Models

In this paper we demonstrate methods for reliable and efficient training...
research
03/29/2022

Autoregressive Co-Training for Learning Discrete Speech Representations

While several self-supervised approaches for learning discrete speech re...
research
04/21/2020

Discrete Variational Attention Models for Language Generation

Variational autoencoders have been widely applied for natural language g...
research
10/16/2018

The LORACs prior for VAEs: Letting the Trees Speak for the Data

In variational autoencoders, the prior on the latent codes z is often tr...
research
05/04/2018

Is Information in the Brain Represented in Continuous or Discrete Form?

The question of continuous-versus-discrete information representation in...

Please sign up or login with your details

Forgot password? Click here to reset