AI-GAN: Attack-Inspired Generation of Adversarial Examples

02/06/2020
by   Tao Bai, et al.
21

Adversarial examples that can fool deep models are mainly crafted by adding small perturbations imperceptible to human eyes. There are various optimization-based methods in the literature to generate adversarial perturbations, most of which are time-consuming. AdvGAN, a method proposed by Xiao et al. in IJCAI 2018, employs Generative Adversarial Networks (GAN) to generate adversarial perturbation with original images as inputs, which is faster than optimization-based methods at inference time. AdvGAN, however, fixes the target classes in the training and we find it difficult to train AdvGAN when it is modified to take original images and target classes as inputs. In this paper, we propose GAN () with a different training strategy to solve this problem. is a two-stage method, in which we use projected gradient descent (PGD) attack to inspire the training of GAN in the first stage and apply standard training of GAN in the second stage. Once trained, the Generator can approximate the conditional distribution of adversarial instances and generate adversarial perturbations given different target classes. We conduct experiments and evaluate the performance of on MNIST and . Compared with AdvGAN, achieves higher attack success rates with similar perturbation magnitudes.

READ FULL TEXT

page 5

page 6

research
04/16/2019

AT-GAN: A Generative Attack Model for Adversarial Transferring on Generative Adversarial Nets

Recent studies have discovered the vulnerability of Deep Neural Networks...
research
11/12/2019

Few-Features Attack to Fool Machine Learning Models through Mask-Based GAN

GAN is a deep-learning based generative approach to generate contents su...
research
05/07/2019

Generating Realistic Unrestricted Adversarial Inputs using Dual-Objective GAN Training

The correctness of deep neural networks is well-known to be vulnerable t...
research
06/20/2021

Attack to Fool and Explain Deep Networks

Deep visual models are susceptible to adversarial perturbations to input...
research
11/19/2021

Towards Efficiently Evaluating the Robustness of Deep Neural Networks in IoT Systems: A GAN-based Method

Intelligent Internet of Things (IoT) systems based on deep neural networ...
research
06/09/2020

GAP++: Learning to generate target-conditioned adversarial examples

Adversarial examples are perturbed inputs which can cause a serious thre...
research
06/27/2018

Customizing an Adversarial Example Generator with Class-Conditional GANs

Adversarial examples are intentionally crafted data with the purpose of ...

Please sign up or login with your details

Forgot password? Click here to reset