Analysis of universal adversarial perturbations

Deep networks have recently been shown to be vulnerable to universal perturbations: there exist very small image-agnostic perturbations that cause most natural images to be misclassified by such classifiers. In this paper, we propose the first quantitative analysis of the robustness of classifiers to universal perturbations, and draw a formal link between the robustness to universal perturbations, and the geometry of the decision boundary. Specifically, we establish theoretical bounds on the robustness of classifiers under two decision boundary models (flat and curved models). We show in particular that the robustness of deep networks to universal perturbations is driven by a key property of their curvature: there exists shared directions along which the decision boundary of deep networks is systematically positively curved. Under such conditions, we prove the existence of small universal perturbations. Our analysis further provides a novel geometric method for computing universal perturbations, in addition to explaining their properties.

READ FULL TEXT

page 4

page 5

page 6

page 7

page 10

research
10/26/2016

Universal adversarial perturbations

Given a state-of-the-art deep neural network classifier, we show the exi...
research
05/26/2017

Classification regions of deep neural networks

The goal of this paper is to analyze the geometric properties of deep ne...
research
11/06/2018

SparseFool: a few pixels make a big difference

Deep Neural Networks have achieved extraordinary results on image classi...
research
02/09/2015

Analysis of classifiers' robustness to adversarial perturbations

The goal of this paper is to analyze an intriguing phenomenon recently d...
research
09/11/2018

On the Structural Sensitivity of Deep Convolutional Networks to the Directions of Fourier Basis Functions

Data-agnostic quasi-imperceptible perturbations on inputs can severely d...
research
08/30/2022

Robustness and invariance properties of image classifiers

Deep neural networks have achieved impressive results in many image clas...
research
08/31/2016

Robustness of classifiers: from adversarial to random noise

Several recent works have shown that state-of-the-art classifiers are vu...

Please sign up or login with your details

Forgot password? Click here to reset