Robust Single-step Adversarial Training with Regularizer

02/05/2021
by   Lehui Xie, et al.
0

High cost of training time caused by multi-step adversarial example generation is a major challenge in adversarial training. Previous methods try to reduce the computational burden of adversarial training using single-step adversarial example generation schemes, which can effectively improve the efficiency but also introduce the problem of catastrophic overfitting, where the robust accuracy against Fast Gradient Sign Method (FGSM) can achieve nearby 100% whereas the robust accuracy against Projected Gradient Descent (PGD) suddenly drops to 0% over a single epoch. To address this problem, we propose a novel Fast Gradient Sign Method with PGD Regularization (FGSMPR) to boost the efficiency of adversarial training without catastrophic overfitting. Our core idea is that single-step adversarial training can not learn robust internal representations of FGSM and PGD adversarial examples. Therefore, we design a PGD regularization term to encourage similar embeddings of FGSM and PGD adversarial examples. The experiments demonstrate that our proposed method can train a robust deep network for L_∞-perturbations with FGSM adversarial training and reduce the gap to multi-step adversarial training.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/05/2020

Understanding Catastrophic Overfitting in Single-step Adversarial Training

Adversarial examples are perturbed inputs that are designed to deceive m...
research
01/12/2020

Fast is better than free: Revisiting adversarial training

Adversarial training, a method for learning robust deep networks, is typ...
research
12/23/2021

Revisiting and Advancing Fast Adversarial Training Through The Lens of Bi-Level Optimization

Adversarial training (AT) has become a widely recognized defense mechani...
research
10/13/2020

Toward Few-step Adversarial Training from a Frequency Perspective

We investigate adversarial-sample generation methods from a frequency do...
research
02/23/2023

Investigating Catastrophic Overfitting in Fast Adversarial Training: A Self-fitting Perspective

Although fast adversarial training provides an efficient approach for bu...
research
09/06/2022

Bag of Tricks for FGSM Adversarial Training

Adversarial training (AT) with samples generated by Fast Gradient Sign M...
research
04/04/2021

Reliably fast adversarial training via latent adversarial perturbation

While multi-step adversarial training is widely popular as an effective ...

Please sign up or login with your details

Forgot password? Click here to reset