Self-Damaging Contrastive Learning

06/06/2021
by   Ziyu Jiang, et al.
0

The recent breakthrough achieved by contrastive learning accelerates the pace for deploying unsupervised training on real-world data applications. However, unlabeled data in reality is commonly imbalanced and shows a long-tail distribution, and it is unclear how robustly the latest contrastive learning methods could perform in the practical scenario. This paper proposes to explicitly tackle this challenge, via a principled framework called Self-Damaging Contrastive Learning (SDCLR), to automatically balance the representation learning without knowing the classes. Our main inspiration is drawn from the recent finding that deep models have difficult-to-memorize samples, and those may be exposed through network pruning. It is further natural to hypothesize that long-tail samples are also tougher for the model to learn well due to insufficient examples. Hence, the key innovation in SDCLR is to create a dynamic self-competitor model to contrast with the target model, which is a pruned version of the latter. During training, contrasting the two models will lead to adaptive online mining of the most easily forgotten samples for the current target model, and implicitly emphasize them more in the contrastive loss. Extensive experiments across multiple datasets and imbalance settings show that SDCLR significantly improves not only overall accuracies but also balancedness, in terms of linear evaluation on the full-shot and few-shot settings. Our code is available at: https://github.com/VITA-Group/SDCLR.

READ FULL TEXT

page 2

page 9

research
11/01/2021

Improving Contrastive Learning on Imbalanced Seed Data via Open-World Sampling

Contrastive learning approaches have achieved great success in learning ...
research
09/06/2022

Robust and Efficient Imbalanced Positive-Unlabeled Learning with Self-supervision

Learning from positive and unlabeled (PU) data is a setting where the le...
research
06/08/2023

On the Effectiveness of Out-of-Distribution Data in Self-Supervised Long-Tail Learning

Though Self-supervised learning (SSL) has been widely studied as a promi...
research
05/23/2022

ImGCL: Revisiting Graph Contrastive Learning on Imbalanced Node Classification

Graph contrastive learning (GCL) has attracted a surge of attention due ...
research
07/12/2023

Contrastive Learning for Conversion Rate Prediction

Conversion rate (CVR) prediction plays an important role in advertising ...
research
07/26/2021

Parametric Contrastive Learning

In this paper, we propose Parametric Contrastive Learning (PaCo) to tack...
research
09/13/2021

Online Unsupervised Learning of Visual Representations and Categories

Real world learning scenarios involve a nonstationary distribution of cl...

Please sign up or login with your details

Forgot password? Click here to reset