Convergence of End-to-End Training in Deep Unsupervised Contrasitive Learning

02/17/2020
by   Zixin Wen, et al.
0

Unsupervised contrastive learning has gained increasing attention in the latest research and has proven to be a powerful method for learning representations from unlabeled data. However, little theoretical analysis was known for this framework. In this paper, we study the optimization of deep unsupervised contrastive learning. We prove that for over-parameterized neural networks, by applying end-to-end training that simultaneously updates two deep neural networks, one can find an approximate stationary solution for the non-convex contrastive loss. This result is inherently different from the existing over-parameterized analysis in the supervised setting because, in contrast to learning a specific target function, unsupervised contrastive learning tries to encode the unlabeled data distribution into the neural networks, which generally has no optimal solution. Our analysis provides theoretical insights into the practical success of these unsupervised pretraining methods.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
01/21/2022

How does unlabeled data improve generalization in self-training? A one-hidden-layer theoretical analysis

Self-training, a semi-supervised learning algorithm, leverages a large a...
research
12/02/2020

About contrastive unsupervised representation learning for classification and its convergence

Contrastive representation learning has been recently proved to be very ...
research
05/19/2023

Towards understanding neural collapse in supervised contrastive learning with the information bottleneck method

Neural collapse describes the geometry of activation in the final layer ...
research
12/21/2021

Max-Margin Contrastive Learning

Standard contrastive learning approaches usually require a large number ...
research
11/27/2017

DeepChess: End-to-End Deep Neural Network for Automatic Learning in Chess

We present an end-to-end learning method for chess, relying on deep neur...
research
10/07/2020

Theoretical Analysis of Self-Training with Deep Networks on Unlabeled Data

Self-training algorithms, which train a model to fit pseudolabels predic...
research
05/07/2019

Contrastive Learning for Lifted Networks

In this work we address supervised learning via lifted network formulati...

Please sign up or login with your details

Forgot password? Click here to reset