Provable Stochastic Optimization for Global Contrastive Learning: Small Batch Does Not Harm Performance

02/24/2022
by   Zhuoning Yuan, et al.
0

In this paper, we study contrastive learning from an optimization perspective, aiming to analyze and address a fundamental issue of existing contrastive learning methods that either rely on a large batch size or a large dictionary. We consider a global objective for contrastive learning, which contrasts each positive pair with all negative pairs for an anchor point. From the optimization perspective, we explain why existing methods such as SimCLR requires a large batch size in order to achieve a satisfactory result. In order to remove such requirement, we propose a memory-efficient Stochastic Optimization algorithm for solving the Global objective of Contrastive Learning of Representations, named SogCLR. We show that its optimization error is negligible under a reasonable condition after a sufficient number of iterations or is diminishing for a slightly different global contrastive objective. Empirically, we demonstrate that on ImageNet with a batch size 256, SogCLR achieves a performance of 69.4 ResNet-50, which is on par with SimCLR (69.3 We also attempt to show that the proposed optimization technique is generic and can be applied to solving other contrastive losses, e.g., two-way contrastive losses for bimodal contrastive learning.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/10/2022

Energy-Based Contrastive Learning of Visual Representations

Contrastive learning is a method of learning visual representations by t...
research
07/12/2023

Mini-Batch Optimization of Contrastive Loss

Contrastive learning has gained significant attention as a method for se...
research
01/18/2021

Scaling Deep Contrastive Learning Batch Size with Almost Constant Peak Memory Usage

Contrastive learning has been applied successfully to learn numerical ve...
research
10/23/2022

Tail Batch Sampling: Approximating Global Contrastive Losses as Optimization over Batch Assignments

Contrastive Learning has recently achieved state-of-the-art performance ...
research
10/13/2021

Decoupled Contrastive Learning

Contrastive learning (CL) is one of the most successful paradigms for se...
research
12/22/2021

Simple and Effective Balance of Contrastive Losses

Contrastive losses have long been a key ingredient of deep metric learni...
research
10/20/2020

BYOL works even without batch statistics

Bootstrap Your Own Latent (BYOL) is a self-supervised learning approach ...

Please sign up or login with your details

Forgot password? Click here to reset