Provable Guarantees for Self-Supervised Deep Learning with Spectral Contrastive Loss

06/08/2021
by   Jeff Z. HaoChen, et al.
2

Recent works in self-supervised learning have advanced the state-of-the-art by relying on the contrastive learning paradigm, which learns representations by pushing positive pairs, or similar examples from the same class, closer together while keeping negative pairs far apart. Despite the empirical successes, theoretical foundations are limited – prior analyses assume conditional independence of the positive pairs given the same class label, but recent empirical applications use heavily correlated positive pairs (i.e., data augmentations of the same image). Our work analyzes contrastive learning without assuming conditional independence of positive pairs using a novel concept of the augmentation graph on data. Edges in this graph connect augmentations of the same data, and ground-truth classes naturally form connected sub-graphs. We propose a loss that performs spectral decomposition on the population augmentation graph and can be succinctly written as a contrastive learning objective on neural net representations. Minimizing this objective leads to features with provable accuracy guarantees under linear probe evaluation. By standard generalization bounds, these accuracy guarantees also hold when minimizing the training contrastive loss. Empirically, the features learned by our objective can match or outperform several strong baselines on benchmark vision datasets. In all, this work provides the first provable analysis for contrastive learning where guarantees for linear probe evaluation can apply to realistic empirical settings.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
07/18/2023

Towards the Sparseness of Projection Head in Self-Supervised Learning

In recent years, self-supervised learning (SSL) has emerged as a promisi...
research
08/09/2023

When and How Does Known Class Help Discover Unknown Ones? Provable Understanding Through Spectral Analysis

Novel Class Discovery (NCD) aims at inferring novel classes in an unlabe...
research
07/05/2022

Features Based Adaptive Augmentation for Graph Contrastive Learning

Self-Supervised learning aims to eliminate the need for expensive annota...
research
03/27/2023

Contrastive Learning Is Spectral Clustering On Similarity Graph

Contrastive learning is a powerful self-supervised learning method, but ...
research
06/01/2023

Understanding Augmentation-based Self-Supervised Representation Learning via RKHS Approximation

Good data augmentation is one of the key factors that lead to the empiri...
research
11/20/2022

Can Single-Pass Contrastive Learning Work for Both Homophilic and Heterophilic Graph?

Existing graph contrastive learning (GCL) typically requires two forward...
research
02/11/2022

Conditional Contrastive Learning with Kernel

Conditional contrastive learning frameworks consider the conditional sam...

Please sign up or login with your details

Forgot password? Click here to reset