Detecting Adversarial Examples by Input Transformations, Defense Perturbations, and Voting

01/27/2021
by   Federico Nesti, et al.
0

Over the last few years, convolutional neural networks (CNNs) have proved to reach super-human performance in visual recognition tasks. However, CNNs can easily be fooled by adversarial examples, i.e., maliciously-crafted images that force the networks to predict an incorrect output while being extremely similar to those for which a correct output is predicted. Regular adversarial examples are not robust to input image transformations, which can then be used to detect whether an adversarial example is presented to the network. Nevertheless, it is still possible to generate adversarial examples that are robust to such transformations. This paper extensively explores the detection of adversarial examples via image transformations and proposes a novel methodology, called defense perturbation, to detect robust adversarial examples with the same input transformations the adversarial examples are robust to. Such a defense perturbation is shown to be an effective counter-measure to robust adversarial examples. Furthermore, multi-network adversarial examples are introduced. This kind of adversarial examples can be used to simultaneously fool multiple networks, which is critical in systems that use network redundancy, such as those based on architectures with majority voting over multiple CNNs. An extensive set of experiments based on state-of-the-art CNNs trained on the Imagenet dataset is finally reported.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
11/25/2019

One Man's Trash is Another Man's Treasure: Resisting Adversarial Examples by Adversarial Examples

Modern image classification systems are often built on deep neural netwo...
research
12/25/2018

Adversarial Feature Genome: a Data Driven Adversarial Examples Recognition Method

Convolutional neural networks (CNNs) are easily spoofed by adversarial e...
research
04/22/2019

Using Videos to Evaluate Image Model Robustness

Human visual systems are robust to a wide range of image transformations...
research
01/04/2022

Towards Understanding and Harnessing the Effect of Image Transformation in Adversarial Detection

Deep neural networks (DNNs) are threatened by adversarial examples. Adve...
research
04/23/2018

VectorDefense: Vectorization as a Defense to Adversarial Examples

Training deep neural networks on images represented as grids of pixels h...
research
08/25/2021

Improving Visual Quality of Unrestricted Adversarial Examples with Wavelet-VAE

Traditional adversarial examples are typically generated by adding pertu...
research
06/17/2020

Adversarial Defense by Latent Style Transformations

Machine learning models have demonstrated vulnerability to adversarial a...

Please sign up or login with your details

Forgot password? Click here to reset