Using Single-Step Adversarial Training to Defend Iterative Adversarial Examples

02/22/2020
by   Guanxiong Liu, et al.
0

Adversarial examples have become one of the largest challenges that machine learning models, especially neural network classifiers, face. These adversarial examples break the assumption of attack-free scenario and fool state-of-the-art (SOTA) classifiers with insignificant perturbations to human. So far, researchers achieved great progress in utilizing adversarial training as a defense. However, the overwhelming computational cost degrades its applicability and little has been done to overcome this issue. Single-Step adversarial training methods have been proposed as computationally viable solutions, however they still fail to defend against iterative adversarial examples. In this work, we first experimentally analyze several different SOTA defense methods against adversarial examples. Then, based on observations from experiments, we propose a novel single-step adversarial training method which can defend against both single-step and iterative adversarial examples. Lastly, through extensive evaluations, we demonstrate that our proposed method outperforms the SOTA single-step and iterative adversarial training defense. Compared with ATDA (single-step method) on CIFAR10 dataset, our proposed method achieves 35.67 time. When compared with methods that use BIM or Madry examples (iterative methods) on CIFAR10 dataset, it saves up to 76.03 than 3.78

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/27/2019

Using Intuition from Empirical Properties to Simplify Adversarial Training Defense

Due to the surprisingly good representation power of complex distributio...
research
04/04/2021

Reliably fast adversarial training via latent adversarial perturbation

While multi-step adversarial training is widely popular as an effective ...
research
11/06/2017

Mitigating adversarial effects through randomization

Convolutional neural networks have demonstrated their powerful ability o...
research
04/17/2019

ZK-GanDef: A GAN based Zero Knowledge Adversarial Training Defense for Neural Networks

Neural Network classifiers have been used successfully in a wide range o...
research
10/03/2020

Efficient Robust Training via Backward Smoothing

Adversarial training is so far the most effective strategy in defending ...
research
02/03/2020

Regularizers for Single-step Adversarial Training

The progress in the last decade has enabled machine learning models to a...
research
07/20/2023

Shared Adversarial Unlearning: Backdoor Mitigation by Unlearning Shared Adversarial Examples

Backdoor attacks are serious security threats to machine learning models...

Please sign up or login with your details

Forgot password? Click here to reset