Adversarial Robustness May Be at Odds With Simplicity

01/02/2019
by   Preetum Nakkiran, et al.
0

Current techniques in machine learning are so far are unable to learn classifiers that are robust to adversarial perturbations. However, they are able to learn non-robust classifiers with very high accuracy, even in the presence of random perturbations. Towards explaining this gap, we highlight the hypothesis that robust classification may require more complex classifiers (i.e. more capacity) than standard classification. In this note, we show that this hypothesis is indeed possible, by giving several theoretical examples of classification tasks and sets of "simple" classifiers for which: (1) There exists a simple classifier with high standard accuracy, and also high accuracy under random ℓ_∞ noise. (2) Any simple classifier is not robust: it must have high adversarial loss with ℓ_∞ perturbations. (3) Robust classification is possible, but only with more complex classifiers (exponentially more complex, in some examples). Moreover, there is a quantitative trade-off between robustness and standard accuracy among simple classifiers. This suggests an alternate explanation of this phenomenon, which appears in practice: the tradeoff may occur not because the classification task inherently requires such a tradeoff (as in [Tsipras-Santurkar-Engstrom-Turner-Madry `18]), but because the structure of our current classifiers imposes such a tradeoff.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/09/2015

Analysis of classifiers' robustness to adversarial perturbations

The goal of this paper is to analyze an intriguing phenomenon recently d...
research
04/30/2018

Adversarially Robust Generalization Requires More Data

Machine learning models are often susceptible to adversarial perturbatio...
research
02/20/2020

A Bayes-Optimal View on Adversarial Examples

The ability to fool modern CNN classifiers with tiny perturbations of th...
research
12/16/2019

Constructing a provably adversarially-robust classifier from a high accuracy one

Modern machine learning models with very high accuracy have been shown t...
research
05/25/2018

Adversarial examples from computational constraints

Why are classifiers in high dimension vulnerable to "adversarial" pertur...
research
01/27/2019

An Information-Theoretic Explanation for the Adversarial Fragility of AI Classifiers

We present a simple hypothesis about a compression property of artificia...
research
05/30/2018

There Is No Free Lunch In Adversarial Robustness (But There Are Unexpected Benefits)

We provide a new understanding of the fundamental nature of adversariall...

Please sign up or login with your details

Forgot password? Click here to reset