Uncovering the Limits of Adversarial Training against Norm-Bounded Adversarial Examples

10/07/2020
by   Sven Gowal, et al.
0

Adversarial training and its variants have become de facto standards for learning robust deep neural networks. In this paper, we explore the landscape around adversarial training in a bid to uncover its limits. We systematically study the effect of different training losses, model sizes, activation functions, the addition of unlabeled data (through pseudo-labeling) and other factors on adversarial robustness. We discover that it is possible to train robust models that go well beyond state-of-the-art results by combining larger models, Swish/SiLU activations and model weight averaging. We demonstrate large improvements on CIFAR-10 and CIFAR-100 against ℓ_∞ and ℓ_2 norm-bounded perturbations of size 8/255 and 128/255, respectively. In the setting with additional unlabeled data, we obtain an accuracy under attack of 65.87 with respect to prior art). Without additional data, we obtain an accuracy under attack of 56.43 without any additional modifications, we obtain an accuracy under attack of 80.45 and of 37.70 CIFAR-100.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/22/2017

Regularizing deep networks using efficient layerwise adversarial training

Adversarial training has been shown to regularize deep neural networks i...
research
05/31/2019

Are Labels Required for Improving Adversarial Robustness?

Recent work has uncovered the interesting (and somewhat surprising) find...
research
11/20/2019

Where is the Bottleneck of Adversarial Learning with Unlabeled Data?

Deep neural networks (DNNs) are incredibly brittle due to adversarial ex...
research
12/22/2020

Self-Progressing Robust Training

Enhancing model robustness under new and even adversarial environments i...
research
09/21/2020

Adversarial Training with Stochastic Weight Average

Adversarial training deep neural networks often experience serious overf...
research
03/02/2021

Fixing Data Augmentation to Improve Adversarial Robustness

Adversarial training suffers from robust overfitting, a phenomenon where...
research
10/21/2019

An Alternative Surrogate Loss for PGD-based Adversarial Testing

Adversarial testing methods based on Projected Gradient Descent (PGD) ar...

Please sign up or login with your details

Forgot password? Click here to reset