Fast is better than free: Revisiting adversarial training

01/12/2020
by   Eric Wong, et al.
0

Adversarial training, a method for learning robust deep networks, is typically assumed to be more expensive than traditional training due to the necessity of constructing adversarial examples via a first-order method like projected gradient decent (PGD). In this paper, we make the surprising discovery that it is possible to train empirically robust models using a much weaker and cheaper adversary, an approach that was previously believed to be ineffective, rendering the method no more costly than standard training in practice. Specifically, we show that adversarial training with the fast gradient sign method (FGSM), when combined with random initialization, is as effective as PGD-based training but has significantly lower cost. Furthermore we show that FGSM adversarial training can be further accelerated by using standard techniques for efficient training of deep networks, allowing us to learn a robust CIFAR10 classifier with 45 ϵ=8/255 in 6 minutes, and a robust ImageNet classifier with 43 accuracy at ϵ=2/255 in 12 hours, in comparison to past work based on "free" adversarial training which took 10 and 50 hours to reach the same respective thresholds. Finally, we identify a failure mode referred to as "catastrophic overfitting" which may have caused previous attempts to use FGSM adversarial training to fail. All code for reproducing the experiments in this paper as well as pretrained model weights are at https://github.com/locuslab/fast_adversarial.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/05/2021

Robust Single-step Adversarial Training with Regularizer

High cost of training time caused by multi-step adversarial example gene...
research
05/15/2020

Initializing Perturbations in Multiple Directions for Fast Adversarial Training

Recent developments in the filed of Deep Learning have demonstrated that...
research
02/26/2020

Overfitting in adversarially robust deep learning

It is common practice in deep learning to use overparameterized networks...
research
05/03/2022

Adversarial Training for High-Stakes Reliability

In the future, powerful AI systems may be deployed in high-stakes settin...
research
09/06/2022

Bag of Tricks for FGSM Adversarial Training

Adversarial training (AT) with samples generated by Fast Gradient Sign M...
research
10/11/2021

Boosting Fast Adversarial Training with Learnable Adversarial Initialization

Adversarial training (AT) has been demonstrated to be effective in impro...
research
07/18/2022

Prior-Guided Adversarial Initialization for Fast Adversarial Training

Fast adversarial training (FAT) effectively improves the efficiency of s...

Please sign up or login with your details

Forgot password? Click here to reset