Neural Bayes: A Generic Parameterization Method for Unsupervised Representation Learning

02/20/2020
by   Devansh Arpit, et al.
36

We introduce a parameterization method called Neural Bayes which allows computing statistical quantities that are in general difficult to compute and opens avenues for formulating new objectives for unsupervised representation learning. Specifically, given an observed random variable x and a latent discrete variable z, we can express p(x|z), p(z|x) and p(z) in closed form in terms of a sufficiently expressive function (Eg. neural network) using our parameterization without restricting the class of these distributions. To demonstrate its usefulness, we develop two independent use cases for this parameterization: 1. Mutual Information Maximization (MIM): MIM has become a popular means for self-supervised representation learning. Neural Bayes allows us to compute mutual information between observed random variables x and latent discrete random variables z in closed form. We use this for learning image representations and show its usefulness on downstream classification tasks. 2. Disjoint Manifold Labeling: Neural Bayes allows us to formulate an objective which can optimally label samples from disjoint manifolds present in the support of a continuous distribution. This can be seen as a specific form of clustering where each disjoint manifold in the support is a separate cluster. We design clustering tasks that obey this formulation and empirically show that the model optimally labels the disjoint manifolds. Our code is available at <https://github.com/salesforce/NeuralBayes>

READ FULL TEXT

page 5

page 13

research
08/20/2018

Learning deep representations by mutual information estimation and maximization

Many popular representation-learning algorithms use training objectives ...
research
11/04/2022

Unsupervised Visual Representation Learning via Mutual Information Regularized Assignment

This paper proposes Mutual Information Regularized Assignment (MIRA), a ...
research
01/19/2023

DiME: Maximizing Mutual Information by a Difference of Matrix-Based Entropies

We introduce an information-theoretic quantity with similar properties t...
research
11/09/2020

Estimating Total Correlation with Mutual Information Bounds

Total correlation (TC) is a fundamental concept in information theory to...
research
06/15/2020

Dissimilarity Mixture Autoencoder for Deep Clustering

In this paper, we introduce the Dissimilarity Mixture Autoencoder (DMAE)...
research
03/13/2020

DHOG: Deep Hierarchical Object Grouping

Recently, a number of competitive methods have tackled unsupervised repr...
research
12/07/2021

Unsupervised Representation Learning via Neural Activation Coding

We present neural activation coding (NAC) as a novel approach for learni...

Please sign up or login with your details

Forgot password? Click here to reset