Rethinking Empirical Evaluation of Adversarial Robustness Using First-Order Attack Methods

06/01/2020
by   Kyungmi Lee, et al.
0

We identify three common cases that lead to overestimation of adversarial accuracy against bounded first-order attack methods, which is popularly used as a proxy for adversarial robustness in empirical studies. For each case, we propose compensation methods that either address sources of inaccurate gradient computation, such as numerical instability near zero and non-differentiability, or reduce the total number of back-propagations for iterative attacks by approximating second-order information. These compensation methods can be combined with existing attack methods for a more precise empirical evaluation metric. We illustrate the impact of these three cases with examples of practical interest, such as benchmarking model capacity and regularization techniques for robustness. Overall, our work shows that overestimated adversarial accuracy that is not indicative of robustness is prevalent even for conventionally trained deep neural networks, and highlights cautions of using empirical evaluation without guaranteed bounds.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
04/04/2020

Adversarial Robustness through Regularization: A Second-Order Approach

Adversarial training is a common approach to improving the robustness of...
research
11/24/2021

Thundernna: a white box adversarial attack

The existing work shows that the neural network trained by naive gradien...
research
09/10/2018

Second-Order Adversarial Attack and Certifiable Robustness

We propose a powerful second-order attack method that outperforms existi...
research
03/09/2020

An Empirical Evaluation on Robustness and Uncertainty of Regularization Methods

Despite apparent human-level performances of deep neural networks (DNN),...
research
11/16/2021

An Empirical Evaluation of the Impact of New York's Bail Reform on Crime Using Synthetic Controls

We conduct an empirical evaluation of the impact of New York's bail refo...
research
09/03/2019

Metric Learning for Adversarial Robustness

Deep networks are well-known to be fragile to adversarial attacks. Using...
research
10/02/2022

Optimization for Robustness Evaluation beyond ℓ_p Metrics

Empirical evaluation of deep learning models against adversarial attacks...

Please sign up or login with your details

Forgot password? Click here to reset