Poisoning and Backdooring Contrastive Learning

06/17/2021
by   Nicholas Carlini, et al.
0

Contrastive learning methods like CLIP train on noisy and uncurated training datasets. This is cheaper than labeling datasets manually, and even improves out-of-distribution robustness. We show that this practice makes backdoor and poisoning attacks a significant threat. By poisoning just 0.005 (e.g., just 150 images of the 3 million-example Conceptual Captions dataset), we can cause the model to misclassify test images by overlaying a small patch. Targeted poisoning attacks, whereby the model misclassifies a particular test input with an adversarially-desired label, are even easier requiring control of less than 0.0001 Our attacks call into question whether training on noisy and uncurated Internet scrapes is desirable.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/22/2022

Indiscriminate Poisoning Attacks on Unsupervised Contrastive Learning

Indiscriminate data poisoning attacks are quite effective against superv...
research
04/03/2018

Poison Frogs! Targeted Clean-Label Poisoning Attacks on Neural Networks

Data poisoning is a type of adversarial attack on machine learning model...
research
03/03/2023

NCL: Textual Backdoor Defense Using Noise-augmented Contrastive Learning

At present, backdoor attacks attract attention as they do great harm to ...
research
05/07/2019

Contrastive Learning for Lifted Networks

In this work we address supervised learning via lifted network formulati...
research
05/03/2022

Data Determines Distributional Robustness in Contrastive Language Image Pre-training (CLIP)

Contrastively trained image-text models such as CLIP, ALIGN, and BASIC h...
research
01/29/2022

Investigating Why Contrastive Learning Benefits Robustness Against Label Noise

Self-supervised contrastive learning has recently been shown to be very ...

Please sign up or login with your details

Forgot password? Click here to reset