Investigating Catastrophic Overfitting in Fast Adversarial Training: A Self-fitting Perspective

02/23/2023
by   Zhengbao He, et al.
1

Although fast adversarial training provides an efficient approach for building robust networks, it may suffer from a serious problem known as catastrophic overfitting (CO), where the multi-step robust accuracy suddenly collapses to zero. In this paper, we for the first time decouple the FGSM examples into data-information and self-information, which reveals an interesting phenomenon called "self-fitting". Self-fitting, i.e., DNNs learn the self-information embedded in single-step perturbations, naturally leads to the occurrence of CO. When self-fitting occurs, the network experiences an obvious "channel differentiation" phenomenon that some convolution channels accounting for recognizing self-information become dominant, while others for data-information are suppressed. In this way, the network learns to only recognize images with sufficient self-information and loses generalization ability to other types of data. Based on self-fitting, we provide new insight into the existing methods to mitigate CO and extend CO to multi-step adversarial training. Our findings reveal a self-learning mechanism in adversarial training and open up new perspectives for suppressing different kinds of information to mitigate CO.

READ FULL TEXT

page 1

page 2

research
06/06/2022

Fast Adversarial Training with Adaptive Step Size

While adversarial training and its variants have shown to be the most ef...
research
02/05/2021

Robust Single-step Adversarial Training with Regularizer

High cost of training time caused by multi-step adversarial example gene...
research
10/11/2022

Stable and Efficient Adversarial Training through Local Linearization

There has been a recent surge in single-step adversarial training as it ...
research
03/29/2021

ZeroGrad : Mitigating and Explaining Catastrophic Overfitting in FGSM Adversarial Training

Making deep neural networks robust to small adversarial noises has recen...
research
11/24/2021

Subspace Adversarial Training

Single-step adversarial training (AT) has received wide attention as it ...
research
02/02/2022

Make Some Noise: Reliable and Efficient Single-Step Adversarial Training

Recently, Wong et al. showed that adversarial training with single-step ...
research
06/16/2022

Catastrophic overfitting is a bug but also a feature

Despite clear computational advantages in building robust neural network...

Please sign up or login with your details

Forgot password? Click here to reset