Controlling Over-generalization and its Effect on Adversarial Examples Generation and Detection

08/21/2018
by   Mahdieh Abbasi, et al.
0

Convolutional Neural Networks (CNNs) allowed improving the state-of-the-art for many vision applications. However, naive CNNs suffer from two serious issues: vulnerability to adversarial examples and making incorrect but confident predictions for out-distribution samples. In this paper, we draw a connection between these two issues of CNNs through over-generalization. We reveal an augmented CNN (an extra output class added) as a simple yet effective end-to-end approach has the capacity for controlling over-generalization. We demonstrate training an augmented CNN on only a properly selected natural out-distribution dataset and interpolated samples empowers it to classify a wide range of unseen out-distribution samples as dustbin. Meanwhile, its misclassification rates on a broad spectrum of well-known black-box adversaries drop drastically as it classifies a portion of adversaries as dustbin class (rejection option) while correctly classifies some of the remaining. However, such an augmented CNN is never trained with any types of adversaries. Finally, generation of white-box adversarial attacks using augmented CNNs can be harder as the attack algorithms have to avoid dustbin regions for generating actual adversaries.

READ FULL TEXT

page 8

page 13

research
05/17/2020

Toward Adversarial Robustness by Diversity in an Ensemble of Specialized Deep Neural Networks

We aim at demonstrating the influence of diversity in the ensemble of CN...
research
02/20/2018

Out-distribution training confers robustness to deep neural networks

The easiness at which adversarial instances can be generated in deep neu...
research
10/18/2019

Toward Metrics for Differentiating Out-of-Distribution Sets

Vanilla CNNs, as uncalibrated classifiers, suffer from classifying out-o...
research
11/18/2020

Adversarial Profiles: Detecting Out-Distribution Adversarial Samples in Pre-trained CNNs

Despite high accuracy of Convolutional Neural Networks (CNNs), they are ...
research
06/14/2020

On the transferability of adversarial examples between convex and 01 loss models

We show that white box adversarial examples do not transfer effectively ...
research
09/07/2022

Securing the Spike: On the Transferabilty and Security of Spiking Neural Networks to Adversarial Examples

Spiking neural networks (SNNs) have attracted much attention for their h...
research
06/14/2018

Copycat CNN: Stealing Knowledge by Persuading Confession with Random Non-Labeled Data

In the past few years, Convolutional Neural Networks (CNNs) have been ac...

Please sign up or login with your details

Forgot password? Click here to reset