Robustness of classifiers: from adversarial to random noise

08/31/2016
by   Alhussein Fawzi, et al.
0

Several recent works have shown that state-of-the-art classifiers are vulnerable to worst-case (i.e., adversarial) perturbations of the datapoints. On the other hand, it has been empirically observed that these same classifiers are relatively robust to random noise. In this paper, we propose to study a semi-random noise regime that generalizes both the random and worst-case noise regimes. We propose the first quantitative analysis of the robustness of nonlinear classifiers in this general noise regime. We establish precise theoretical bounds on the robustness of classifiers in this general regime, which depend on the curvature of the classifier's decision boundary. Our bounds confirm and quantify the empirical observations that classifiers satisfying curvature constraints are robust to random noise. Moreover, we quantify the robustness of classifiers in terms of the subspace dimension in the semi-random noise regime, and show that our bounds remarkably interpolate between the worst-case and random noise regimes. We perform experiments and show that the derived bounds provide very accurate estimates when applied to various state-of-the-art deep neural networks and datasets. This result suggests bounds on the curvature of the classifiers' decision boundaries that we support experimentally, and more generally offers important insights onto the geometry of high dimensional classification problems.

READ FULL TEXT

page 5

page 6

page 8

page 9

page 10

page 15

page 17

research
02/22/2018

Robustness of classifiers to uniform ℓ_p and Gaussian noise

We study the robustness of classifiers to various kinds of random noise ...
research
02/09/2015

Analysis of classifiers' robustness to adversarial perturbations

The goal of this paper is to analyze an intriguing phenomenon recently d...
research
05/26/2017

Analysis of universal adversarial perturbations

Deep networks have recently been shown to be vulnerable to universal per...
research
06/07/2020

Consistency Regularization for Certified Robustness of Smoothed Classifiers

A recent technique of randomized smoothing has shown that the worst-case...
research
05/26/2017

Classification regions of deep neural networks

The goal of this paper is to analyze the geometric properties of deep ne...
research
02/22/2021

On the robustness of randomized classifiers to adversarial examples

This paper investigates the theory of robustness against adversarial att...
research
10/14/2020

Linking average- and worst-case perturbation robustness via class selectivity and dimensionality

Representational sparsity is known to affect robustness to input perturb...

Please sign up or login with your details

Forgot password? Click here to reset