Generating Realistic Unrestricted Adversarial Inputs using Dual-Objective GAN Training

05/07/2019
by   Isaac Dunn, et al.
0

The correctness of deep neural networks is well-known to be vulnerable to small, 'adversarial' perturbations of their inputs. Although studying these attacks is valuable, they do not necessarily conform to any real-world threat model. This has led to interest in the generation of (and robustness to) unrestricted adversarial inputs, which are not constructed as small perturbations of correctly-classified ground-truth inputs. We introduce a novel algorithm to generate realistic unrestricted adversarial inputs, in the sense that they cannot reliably be distinguished from the training dataset by a human. This is achieved by modifying generative adversarial networks: a generator neural network is trained to construct examples that deceive a fixed target network (so they are adversarial) while also deceiving the usual co-training discriminator network (so they are realistic). Our approach is demonstrated by the generation of unrestricted adversarial inputs for a trained image classifier that is robust to perturbation-based attacks. We find that human judges are unable to identify which image out of ten was generated by our method about 50 percent of the time, providing evidence that they are moderately realistic.

READ FULL TEXT

page 19

page 20

page 21

page 22

page 23

page 24

page 25

page 27

research
05/09/2017

Generative Adversarial Trainer: Defense to Adversarial Perturbations with GAN

We propose a novel technique to make neural network robust to adversaria...
research
02/06/2020

AI-GAN: Attack-Inspired Generation of Adversarial Examples

Adversarial examples that can fool deep models are mainly crafted by add...
research
12/29/2021

Repairing Adversarial Texts through Perturbation

It is known that neural networks are subject to attacks through adversar...
research
12/05/2019

Label-Consistent Backdoor Attacks

Deep neural networks have been demonstrated to be vulnerable to backdoor...
research
04/19/2022

CorrGAN: Input Transformation Technique Against Natural Corruptions

Because of the increasing accuracy of Deep Neural Networks (DNNs) on dif...
research
09/12/2023

Using Reed-Muller Codes for Classification with Rejection and Recovery

When deploying classifiers in the real world, users expect them to respo...
research
02/18/2021

Random Projections for Improved Adversarial Robustness

We propose two training techniques for improving the robustness of Neura...

Please sign up or login with your details

Forgot password? Click here to reset