Verifying the Causes of Adversarial Examples

10/19/2020
by   Honglin Li, et al.
0

The robustness of neural networks is challenged by adversarial examples that contain almost imperceptible perturbations to inputs, which mislead a classifier to incorrect outputs in high confidence. Limited by the extreme difficulty in examining a high-dimensional image space thoroughly, research on explaining and justifying the causes of adversarial examples falls behind studies on attacks and defenses. In this paper, we present a collection of potential causes of adversarial examples and verify (or partially verify) them through carefully-designed controlled experiments. The major causes of adversarial examples include model linearity, one-sum constraint, and geometry of the categories. To control the effect of those causes, multiple techniques are applied such as L_2 normalization, replacement of loss functions, construction of reference datasets, and novel models using multi-layer perceptron probabilistic neural networks (MLP-PNN) and density estimation (DE). Our experiment results show that geometric factors tend to be more direct causes and statistical factors magnify the phenomenon, especially for assigning high prediction confidence. We believe this paper will inspire more studies to rigorously investigate the root causes of adversarial examples, which in turn provide useful guidance on designing more robust models.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
01/10/2019

Image Transformation can make Neural Networks more robust against Adversarial Examples

Neural networks are being applied in many tasks related to IoT with enco...
research
05/20/2017

Adversarial Examples Are Not Easily Detected: Bypassing Ten Detection Methods

Neural networks are known to be vulnerable to adversarial examples: inpu...
research
12/20/2014

Explaining and Harnessing Adversarial Examples

Several machine learning models, including neural networks, consistently...
research
02/13/2018

Predicting Adversarial Examples with High Confidence

It has been suggested that adversarial examples cause deep learning mode...
research
07/12/2022

Exploring Adversarial Examples and Adversarial Robustness of Convolutional Neural Networks by Mutual Information

A counter-intuitive property of convolutional neural networks (CNNs) is ...
research
05/26/2022

An Analytic Framework for Robust Training of Artificial Neural Networks

The reliability of a learning model is key to the successful deployment ...
research
06/23/2021

Adversarial Examples in Multi-Layer Random ReLU Networks

We consider the phenomenon of adversarial examples in ReLU networks with...

Please sign up or login with your details

Forgot password? Click here to reset