An Alternative Surrogate Loss for PGD-based Adversarial Testing

10/21/2019
by   Sven Gowal, et al.
8

Adversarial testing methods based on Projected Gradient Descent (PGD) are widely used for searching norm-bounded perturbations that cause the inputs of neural networks to be misclassified. This paper takes a deeper look at these methods and explains the effect of different hyperparameters (i.e., optimizer, step size and surrogate loss). We introduce the concept of MultiTargeted testing, which makes clever use of alternative surrogate losses, and explain when and how MultiTargeted is guaranteed to find optimal perturbations. Finally, we demonstrate that MultiTargeted outperforms more sophisticated methods and often requires less iterative steps than other variants of PGD found in the literature. Notably, MultiTargeted ranks first on MadryLab's white-box MNIST and CIFAR-10 leaderboards, reducing the accuracy of their MNIST model to 88.36 accuracy of their CIFAR-10 model to 44.03 MultiTargeted also ranks first on the TRADES leaderboard reducing the accuracy of their CIFAR-10 model to 53.07 ϵ = 0.031).

READ FULL TEXT

page 12

page 14

page 15

research
09/18/2023

Reducing Adversarial Training Cost with Gradient Approximation

Deep learning models have achieved state-of-the-art performances in vari...
research
11/18/2019

WITCHcraft: Efficient PGD attacks with random step size

State-of-the-art adversarial attacks on neural networks use expensive it...
research
12/06/2018

Max-Margin Adversarial (MMA) Training: Direct Input Space Margin Maximization through Adversarial Training

We propose Max-Margin Adversarial (MMA) training for directly maximizing...
research
08/16/2018

Distributionally Adversarial Attack

Recent work on adversarial attack has shown that Projected Gradient Desc...
research
10/07/2020

Uncovering the Limits of Adversarial Training against Norm-Bounded Adversarial Examples

Adversarial training and its variants have become de facto standards for...
research
02/21/2020

Adversarial Detection and Correction by Matching Prediction Distributions

We present a novel adversarial detection and correction method for machi...
research
09/14/2020

Deforming the Loss Surface to Affect the Behaviour of the Optimizer

In deep learning, it is usually assumed that the optimization process is...

Please sign up or login with your details

Forgot password? Click here to reset