On the Effectiveness of Adversarial Training against Backdoor Attacks

02/22/2022
by   Yinghua Gao, et al.
1

DNNs' demand for massive data forces practitioners to collect data from the Internet without careful check due to the unacceptable cost, which brings potential risks of backdoor attacks. A backdoored model always predicts a target class in the presence of a predefined trigger pattern, which can be easily realized via poisoning a small amount of data. In general, adversarial training is believed to defend against backdoor attacks since it helps models to keep their prediction unchanged even if we perturb the input image (as long as within a feasible range). Unfortunately, few previous studies succeed in doing so. To explore whether adversarial training could defend against backdoor attacks or not, we conduct extensive experiments across different threat models and perturbation budgets, and find the threat model in adversarial training matters. For instance, adversarial training with spatial adversarial examples provides notable robustness against commonly-used patch-based backdoor attacks. We further propose a hybrid strategy which provides satisfactory robustness across different backdoor attacks.

READ FULL TEXT
research
10/14/2019

Confidence-Calibrated Adversarial Training: Towards Robust Models Generalizing Beyond the Attack Used During Training

Adversarial training is the standard to train models robust against adve...
research
02/16/2023

Masking and Mixing Adversarial Training

While convolutional neural networks (CNNs) have achieved excellent perfo...
research
06/12/2023

How robust accuracy suffers from certified training with convex relaxations

Adversarial attacks pose significant threats to deploying state-of-the-a...
research
02/26/2021

What Doesn't Kill You Makes You Robust(er): Adversarial Training against Poisons and Backdoors

Data poisoning is a threat model in which a malicious actor tampers with...
research
03/17/2022

On the Properties of Adversarially-Trained CNNs

Adversarial Training has proved to be an effective training paradigm to ...
research
09/13/2022

Certified Defences Against Adversarial Patch Attacks on Semantic Segmentation

Adversarial patch attacks are an emerging security threat for real world...
research
05/27/2022

Semi-supervised Semantics-guided Adversarial Training for Trajectory Prediction

Predicting the trajectories of surrounding objects is a critical task in...

Please sign up or login with your details

Forgot password? Click here to reset