Target Training Does Adversarial Training Without Adversarial Samples

02/09/2021
by   Blerta Lindqvist, et al.
0

Neural network classifiers are vulnerable to misclassification of adversarial samples, for which the current best defense trains classifiers with adversarial samples. However, adversarial samples are not optimal for steering attack convergence, based on the minimization at the core of adversarial attacks. The minimization perturbation term can be minimized towards 0 by replacing adversarial samples in training with duplicated original samples, labeled differently only for training. Using only original samples, Target Training eliminates the need to generate adversarial samples for training against all attacks that minimize perturbation. In low-capacity classifiers and without using adversarial samples, Target Training exceeds both default CIFAR10 accuracy (84.3 against CW-L_2(κ=0) attack, and 86.6 adversarial samples against attacks that do not minimize perturbation, Target Training exceeds current best defense (69.1 CW-L_2(κ=40) in CIFAR10.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/08/2020

Tricking Adversarial Attacks To Fail

Recent adversarial defense approaches have failed. Untargeted gradient-b...
research
02/05/2021

Optimal Transport as a Defense Against Adversarial Attacks

Deep learning classifiers are now known to have flaws in the representat...
research
08/10/2020

FireBERT: Hardening BERT-based classifiers against adversarial attack

We present FireBERT, a set of three proof-of-concept NLP classifiers har...
research
01/06/2021

DeepPoison: Feature Transfer Based Stealthy Poisoning Attack

Deep neural networks are susceptible to poisoning attacks by purposely p...
research
12/13/2019

Potential adversarial samples for white-box attacks

Deep convolutional neural networks can be highly vulnerable to small per...
research
08/10/2023

Symmetry Defense Against XGBoost Adversarial Perturbation Attacks

We examine whether symmetry can be used to defend tree-based ensemble cl...
research
06/10/2020

Adversarial Attacks on Brain-Inspired Hyperdimensional Computing-Based Classifiers

Being an emerging class of in-memory computing architecture, brain-inspi...

Please sign up or login with your details

Forgot password? Click here to reset