Fast Adversarial Training with Smooth Convergence

08/24/2023
by   Mengnan Zhao, et al.
0

Fast adversarial training (FAT) is beneficial for improving the adversarial robustness of neural networks. However, previous FAT work has encountered a significant issue known as catastrophic overfitting when dealing with large perturbation budgets, the adversarial robustness of models declines to near zero during training. To address this, we analyze the training process of prior FAT work and observe that catastrophic overfitting is accompanied by the appearance of loss convergence outliers. Therefore, we argue a moderately smooth loss convergence process will be a stable FAT process that solves catastrophic overfitting. To obtain a smooth loss convergence process, we propose a novel oscillatory constraint (dubbed ConvergeSmooth) to limit the loss difference between adjacent epochs. The convergence stride of ConvergeSmooth is introduced to balance convergence and smoothing. Likewise, we design weight centralization without introducing additional hyperparameters other than the loss balance coefficient. Our proposed methods are attack-agnostic and thus can improve the training stability of various FAT techniques. Extensive experiments on popular datasets show that the proposed methods efficiently avoid catastrophic overfitting and outperform all previous FAT methods. Code is available at <https://github.com/FAT-CS/ConvergeSmooth>.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
12/09/2022

Understanding and Combating Robust Overfitting via Input Loss Landscape Analysis and Regularization

Adversarial training is widely used to improve the robustness of deep ne...
research
10/11/2022

Stable and Efficient Adversarial Training through Local Linearization

There has been a recent surge in single-step adversarial training as it ...
research
07/06/2020

Understanding and Improving Fast Adversarial Training

A recent line of work focused on making adversarial training computation...
research
07/18/2022

Prior-Guided Adversarial Initialization for Fast Adversarial Training

Fast adversarial training (FAT) effectively improves the efficiency of s...
research
04/01/2023

Improving Fast Adversarial Training with Prior-Guided Knowledge

Fast adversarial training (FAT) is an efficient method to improve robust...
research
09/06/2022

Bag of Tricks for FGSM Adversarial Training

Adversarial training (AT) with samples generated by Fast Gradient Sign M...
research
06/16/2022

Catastrophic overfitting is a bug but also a feature

Despite clear computational advantages in building robust neural network...

Please sign up or login with your details

Forgot password? Click here to reset