Compressive Visual Representations

09/27/2021
by   Kuang-Huei Lee, et al.
6

Learning effective visual representations that generalize well without human supervision is a fundamental problem in order to apply Machine Learning to a wide variety of tasks. Recently, two families of self-supervised methods, contrastive learning and latent bootstrapping, exemplified by SimCLR and BYOL respectively, have made significant progress. In this work, we hypothesize that adding explicit information compression to these algorithms yields better and more robust representations. We verify this by developing SimCLR and BYOL formulations compatible with the Conditional Entropy Bottleneck (CEB) objective, allowing us to both measure and control the amount of compression in the learned representation, and observe their impact on downstream tasks. Furthermore, we explore the relationship between Lipschitz continuity and compression, showing a tractable lower bound on the Lipschitz constant of the encoders we learn. As Lipschitz continuity is closely related to robustness, this provides a new explanation for why compressed models are more robust. Our experiments confirm that adding compression to SimCLR and BYOL significantly improves linear evaluation accuracies and model robustness across a wide range of domain shifts. In particular, the compressed version of BYOL achieves 76.0 Top-1 linear evaluation accuracy on ImageNet with ResNet-50, and 78.8 ResNet-50 2x.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/01/2021

Stochastic Contrastive Learning

While state-of-the-art contrastive Self-Supervised Learning (SSL) models...
research
06/28/2020

Video Representation Learning with Visual Tempo Consistency

Visual tempo, which describes how fast an action goes, has shown its pot...
research
02/24/2023

Generalization Analysis for Contrastive Representation Learning

Recently, contrastive learning has found impressive success in advancing...
research
02/02/2023

Hyperbolic Contrastive Learning

Learning good image representations that are beneficial to downstream ta...
research
07/13/2022

Lipschitz Continuity Retained Binary Neural Network

Relying on the premise that the performance of a binary neural network c...
research
09/03/2019

On the Downstream Performance of Compressed Word Embeddings

Compressing word embeddings is important for deploying NLP models in mem...
research
03/30/2023

Learning in Factored Domains with Information-Constrained Visual Representations

Humans learn quickly even in tasks that contain complex visual informati...

Please sign up or login with your details

Forgot password? Click here to reset