Fast and Stable Adversarial Training through Noise Injection

02/24/2020
by   Leo Schwinn, et al.
0

Adversarial training is the most successful empirical method, to increase the robustness of neural networks against adversarial attacks yet. Unfortunately, this higher robustness is accompanied by considerably higher computational complexity. To date, only adversarial training with expensive multi-step adversarial attacks like Projected Gradient Descent (PGD) proved effective against equally strong attacks. In this paper, we present two ideas that combined enable adversarial training with the computationally less expensive Fast Gradient Sign Method (FGSM). First, we add uniform noise to the initial data point of the FGSM attack, which creates a wider variety of stronger adversaries. Further, we add a learnable regularization step prior to the neural network called Stochastic Augmentation Layer (SAL). Inputs propagated trough the SAL are resampled from a Gaussian distribution. The randomness of the resampling at inference time makes it more complicated for the attacker to construct an adversarial example since the outcome of the model is not known in advance. We show that noise injection in conjunction with FGSM adversarial training achieves comparable results to adversarial training with PGD while being orders of magnitude faster. Moreover, we show superior results in comparison to PGD-based training when combining noise injection and SAL.

READ FULL TEXT

page 2

page 4

research
10/25/2019

Label Smoothing and Logit Squeezing: A Replacement for Adversarial Training?

Adversarial training is one of the strongest defenses against adversaria...
research
06/30/2021

Single-Step Adversarial Training for Semantic Segmentation

Even though deep neural networks succeed on many different tasks includi...
research
06/04/2020

Towards Understanding Fast Adversarial Training

Current neural-network-based classifiers are susceptible to adversarial ...
research
04/18/2020

Single-step Adversarial training with Dropout Scheduling

Deep learning models have shown impressive performance across a spectrum...
research
11/18/2019

WITCHcraft: Efficient PGD attacks with random step size

State-of-the-art adversarial attacks on neural networks use expensive it...
research
12/23/2021

Revisiting and Advancing Fast Adversarial Training Through The Lens of Bi-Level Optimization

Adversarial training (AT) has become a widely recognized defense mechani...
research
09/10/2020

Quantifying the Preferential Direction of the Model Gradient in Adversarial Training With Projected Gradient Descent

Adversarial training, especially projected gradient descent (PGD), has b...

Please sign up or login with your details

Forgot password? Click here to reset