Disentangling Adversarial Robustness and Generalization

12/03/2018
by   David Stutz, et al.
0

Obtaining deep networks that are robust against adversarial examples and generalize well is an open problem. A recent hypothesis even states that both robust and accurate models are impossible, i.e., adversarial robustness and generalization are conflicting goals. In an effort to clarify the relationship between robustness and generalization, we assume an underlying, low-dimensional data manifold and show that: 1. regular adversarial examples leave the manifold; 2. adversarial examples constrained to the manifold, i.e., on-manifold adversarial examples, exist; 3. on-manifold adversarial examples are generalization errors, and on-manifold adversarial training boosts generalization; 4. and regular robustness is independent of generalization. These assumptions imply that both robust and accurate models are possible. However, different models (architectures, training strategies etc.) can exhibit different robustness and generalization characteristics. To confirm our claims, we present extensive experiments on synthetic data (with access to the true manifold) as well as on EMNIST, Fashion-MNIST and CelebA.

READ FULL TEXT

page 1

page 3

page 12

page 14

page 17

page 20

research
10/02/2022

Understanding Adversarial Robustness Against On-manifold Adversarial Examples

Deep neural networks (DNNs) are shown to be vulnerable to adversarial ex...
research
06/13/2018

Manifold Mixup: Encouraging Meaningful On-Manifold Interpolation as a Regularizer

Deep networks often perform well on the data manifold on which they are ...
research
03/04/2021

Hard-label Manifolds: Unexpected Advantages of Query Efficiency for Finding On-manifold Adversarial Examples

Designing deep networks robust to adversarial examples remains an open p...
research
05/02/2019

Adversarial Training with Voronoi Constraints

Adversarial examples are a pervasive phenomenon of machine learning mode...
research
03/23/2018

Generalizability vs. Robustness: Adversarial Examples for Medical Imaging

In this paper, for the first time, we propose an evaluation method for d...
research
04/25/2022

When adversarial examples are excusable

Neural networks work remarkably well in practice and theoretically they ...
research
04/22/2019

Using Videos to Evaluate Image Model Robustness

Human visual systems are robust to a wide range of image transformations...

Please sign up or login with your details

Forgot password? Click here to reset