Towards Understanding How Self-training Tolerates Data Backdoor Poisoning

01/20/2023
by   Soumyadeep Pal, et al.
7

Recent studies on backdoor attacks in model training have shown that polluting a small portion of training data is sufficient to produce incorrect manipulated predictions on poisoned test-time data while maintaining high clean accuracy in downstream tasks. The stealthiness of backdoor attacks has imposed tremendous defense challenges in today's machine learning paradigm. In this paper, we explore the potential of self-training via additional unlabeled data for mitigating backdoor attacks. We begin by making a pilot study to show that vanilla self-training is not effective in backdoor mitigation. Spurred by that, we propose to defend the backdoor attacks by leveraging strong but proper data augmentations in the self-training pseudo-labeling stage. We find that the new self-training regime help in defending against backdoor attacks to a great extent. Its effectiveness is demonstrated through experiments for different backdoor triggers on CIFAR-10 and a combination of CIFAR-10 with an additional unlabeled 500K TinyImages dataset. Finally, we explore the direction of combining self-supervised representation learning with self-training for further improvement in backdoor defense.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
04/04/2023

Defending Against Patch-based Backdoor Attacks on Self-Supervised Learning

Recently, self-supervised learning (SSL) was shown to be vulnerable to p...
research
10/22/2019

Class Mean Vectors, Self Monitoring and Self Learning for Neural Classifiers

In this paper we explore the role of sample mean in building a neural ne...
research
05/21/2021

Backdoor Attacks on Self-Supervised Learning

Large-scale unlabeled data has allowed recent progress in self-supervise...
research
10/13/2022

Demystifying Self-supervised Trojan Attacks

As an emerging machine learning paradigm, self-supervised learning (SSL)...
research
09/16/2022

Dataset Inference for Self-Supervised Models

Self-supervised models are increasingly prevalent in machine learning (M...
research
09/30/2019

Revisiting Self-Training for Neural Sequence Generation

Self-training is one of the earliest and simplest semi-supervised method...
research
10/24/2021

Understanding the World Through Action

The recent history of machine learning research has taught us that machi...

Please sign up or login with your details

Forgot password? Click here to reset