Rethinking Positive Sampling for Contrastive Learning with Kernel

06/03/2022
by   Benoit Dufumier, et al.
0

Data augmentation is a crucial component in unsupervised contrastive learning (CL). It determines how positive samples are defined and, ultimately, the quality of the representation. While efficient augmentations have been found for standard vision datasets, such as ImageNet, it is still an open problem in other applications, such as medical imaging, or in datasets with easy-to-learn but irrelevant imaging features. In this work, we propose a new way to define positive samples using kernel theory along with a novel loss called decoupled uniformity. We propose to integrate prior information, learnt from generative models or given as auxiliary attributes, into contrastive learning, to make it less dependent on data augmentation. We draw a connection between contrastive learning and the conditional mean embedding theory to derive tight bounds on the downstream classification loss. In an unsupervised setting, we empirically demonstrate that CL benefits from generative models, such as VAE and GAN, to less rely on data augmentations. We validate our framework on vision datasets including CIFAR10, CIFAR100, STL10 and ImageNet100 and a brain MRI dataset. In the weakly supervised setting, we demonstrate that our formulation provides state-of-the-art results.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/14/2022

Adversarial Graph Contrastive Learning with Information Regularization

Contrastive learning is an effective unsupervised method in graph repres...
research
02/11/2022

Conditional Contrastive Learning with Kernel

Conditional contrastive learning frameworks consider the conditional sam...
research
10/06/2021

Sharp Learning Bounds for Contrastive Unsupervised Representation Learning

Contrastive unsupervised representation learning (CURL) encourages data ...
research
03/16/2023

Instance-Conditioned GAN Data Augmentation for Representation Learning

Data augmentation has become a crucial component to train state-of-the-a...
research
11/10/2022

Unbiased Supervised Contrastive Learning

Many datasets are biased, namely they contain easy-to-learn features tha...
research
03/20/2022

Partitioning Image Representation in Contrastive Learning

In contrastive learning in the image domain, the anchor and positive sam...
research
03/30/2021

Contrastive Learning of Single-Cell Phenotypic Representations for Treatment Classification

Learning robust representations to discriminate cell phenotypes based on...

Please sign up or login with your details

Forgot password? Click here to reset