Demystifying Self-Supervised Learning: An Information-Theoretical Framework

06/10/2020
by   Yao-Hung Hubert Tsai, et al.
22

Self-supervised representation learning adopts self-defined signals as supervision and uses the learned representation for downstream tasks, such as masked language modeling (e.g., BERT) for natural language processing and contrastive visual representation learning (e.g., SimCLR) for computer vision applications. In this paper, we present a theoretical framework explaining that self-supervised learning is likely to work under the assumption that only the shared information (e.g., contextual information or content) between the input (e.g., non-masked words or original images) and self-supervised signals (e.g., masked-words or augmented images) contributes to downstream tasks. Under this assumption, we demonstrate that self-supervisedly learned representation can extract task-relevant and discard task-irrelevant information. We further connect our theoretical analysis to popular contrastive and predictive (self-supervised) learning objectives. In the experimental section, we provide controlled experiments on two popular tasks: 1) visual representation learning with various self-supervised learning objectives to empirically support our analysis; and 2) visual-textual representation learning to challenge that input and self-supervised signal lie in different modalities.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/15/2020

Self-supervised Learning: Generative or Contrastive

Deep supervised learning has achieved great success in the last decade. ...
research
06/05/2021

Conditional Contrastive Learning: Removing Undesirable Information in Self-Supervised Representations

Self-supervised learning is a form of unsupervised learning that leverag...
research
11/07/2022

On minimal variations for unsupervised representation learning

Unsupervised representation learning aims at describing raw data efficie...
research
08/24/2020

Contrastive learning, multi-view redundancy, and linear models

Self-supervised learning is an empirically successful approach to unsupe...
research
03/21/2021

Self-supervised Representation Learning with Relative Predictive Coding

This paper introduces Relative Predictive Coding (RPC), a new contrastiv...
research
02/13/2021

Understanding Negative Samples in Instance Discriminative Self-supervised Representation Learning

Instance discriminative self-supervised representation learning has been...
research
06/21/2021

Visual Probing: Cognitive Framework for Explaining Self-Supervised Image Representations

Recently introduced self-supervised methods for image representation lea...

Please sign up or login with your details

Forgot password? Click here to reset