Diffusion Models for Imperceptible and Transferable Adversarial Attack

05/14/2023
by   Jianqi Chen, et al.
0

Many existing adversarial attacks generate L_p-norm perturbations on image RGB space. Despite some achievements in transferability and attack success rate, the crafted adversarial examples are easily perceived by human eyes. Towards visual imperceptibility, some recent works explore unrestricted attacks without L_p-norm constraints, yet lacking transferability of attacking black-box models. In this work, we propose a novel imperceptible and transferable attack by leveraging both the generative and discriminative power of diffusion models. Specifically, instead of direct manipulation in pixel space, we craft perturbations in latent space of diffusion models. Combined with well-designed content-preserving structures, we can generate human-insensitive perturbations embedded with semantic clues. For better transferability, we further "deceive" the diffusion model which can be viewed as an additional recognition surrogate, by distracting its attention away from the target regions. To our knowledge, our proposed method, DiffAttack, is the first that introduces diffusion models into adversarial attack field. Extensive experiments on various model structures (including CNNs, Transformers, MLPs) and defense methods have demonstrated our superiority over other attack methods.

READ FULL TEXT

page 2

page 5

page 8

page 17

page 20

page 21

research
09/14/2023

Semantic Adversarial Attacks via Diffusion Models

Traditional adversarial attacks concentrate on manipulating clean exampl...
research
05/22/2023

Latent Magic: An Investigation into Adversarial Examples Crafted in the Semantic Latent Space

Adversarial attacks against Deep Neural Networks(DNN) have been a crutia...
research
07/04/2023

LEAT: Towards Robust Deepfake Disruption in Real-World Scenarios via Latent Ensemble Attack

Deepfakes, malicious visual contents created by generative models, pose ...
research
09/04/2023

Improving Visual Quality and Transferability of Adversarial Attacks on Face Recognition Simultaneously with Adversarial Restoration

Adversarial face examples possess two critical properties: Visual Qualit...
research
05/23/2023

DiffProtect: Generate Adversarial Examples with Diffusion Models for Facial Privacy Protection

The increasingly pervasive facial recognition (FR) systems raise serious...
research
10/13/2022

Adv-Attribute: Inconspicuous and Transferable Adversarial Attack on Face Recognition

Deep learning models have shown their vulnerability when dealing with ad...
research
07/24/2023

Data-free Black-box Attack based on Diffusion Model

Since the training data for the target model in a data-free black-box at...

Please sign up or login with your details

Forgot password? Click here to reset