Unsupervised pre-training helps to conserve views from input distribution

05/30/2019
by   Nicolas Pinchaud, et al.
0

We investigate the effects of the unsupervised pre-training method under the perspective of information theory. If the input distribution displays multiple views of the supervision, then unsupervised pre-training allows to learn hierarchical representation which communicates these views across layers, while disentangling the supervision. Disentanglement of supervision leads learned features to be independent conditionally to the label. In case of binary features, we show that conditional independence allows to extract label's information with a linear model and therefore helps to solve under-fitting. We suppose that representations displaying multiple views help to solve over-fitting because each view provides information that helps to reduce model's variance. We propose a practical method to measure both disentanglement of supervision and quantity of views within a binary representation. We show that unsupervised pre-training helps to conserve views from input distribution, whereas representations learned using supervised models disregard most of them.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/20/2020

What makes for good views for contrastive learning

Contrastive learning between multiple views of the data has recently ach...
research
07/19/2022

Self-Supervision Can Be a Good Few-Shot Learner

Existing few-shot learning (FSL) methods rely on training with a large l...
research
12/20/2014

An Analysis of Unsupervised Pre-training in Light of Recent Advances

Convolutional neural networks perform well on object recognition because...
research
09/27/2020

Unsupervised Pre-training for Biomedical Question Answering

We explore the suitability of unsupervised representation learning metho...
research
04/28/2023

ViP-NeRF: Visibility Prior for Sparse Input Neural Radiance Fields

Neural radiance fields (NeRF) have achieved impressive performances in v...
research
06/01/2023

On Masked Pre-training and the Marginal Likelihood

Masked pre-training removes random input dimensions and learns a model t...
research
10/14/2020

Viewmaker Networks: Learning Views for Unsupervised Representation Learning

Many recent methods for unsupervised representation learning involve tra...

Please sign up or login with your details

Forgot password? Click here to reset