Latent Magic: An Investigation into Adversarial Examples Crafted in the Semantic Latent Space

05/22/2023
by   BoYang Zheng, et al.
0

Adversarial attacks against Deep Neural Networks(DNN) have been a crutial topic ever since <cit.> purposed the vulnerability of DNNs. However, most prior works craft adversarial examples in the pixel space, following the l_p norm constraint. In this paper, we give intuitional explain about why crafting adversarial examples in the latent space is equally efficient and important. We purpose a framework for crafting adversarial examples in semantic latent space based on an pre-trained Variational Auto Encoder from state-of-art Stable Diffusion Model<cit.>. We also show that adversarial examples crafted in the latent space can also achieve a high level of fool rate. However, examples crafted from latent space are often hard to evaluated, as they doesn't follow a certain l_p norm constraint, which is a big challenge for existing researches. To efficiently and accurately evaluate the adversarial examples crafted in the latent space, we purpose a novel evaluation matric based on SSIM<cit.> loss and fool rate.Additionally, we explain why FID<cit.> is not suitable for measuring such adversarial examples. To the best of our knowledge, it's the first evaluation metrics that is specifically designed to evaluate the quality of a adversarial attack. We also investigate the transferability of adversarial examples crafted in the latent space and show that they have superiority over adversarial examples crafted in the pixel space.

READ FULL TEXT

page 2

page 3

page 4

page 6

page 10

research
05/14/2023

Diffusion Models for Imperceptible and Transferable Adversarial Attack

Many existing adversarial attacks generate L_p-norm perturbations on ima...
research
12/09/2020

Generating Out of Distribution Adversarial Attack using Latent Space Poisoning

Traditional adversarial attacks rely upon the perturbations generated by...
research
08/25/2021

Adversarially Robust One-class Novelty Detection

One-class novelty detectors are trained with examples of a particular cl...
research
12/24/2020

Exploring Adversarial Examples via Invertible Neural Networks

Adversarial examples (AEs) are images that can mislead deep neural netwo...
research
09/14/2023

Semantic Adversarial Attacks via Diffusion Models

Traditional adversarial attacks concentrate on manipulating clean exampl...
research
03/29/2023

Latent Feature Relation Consistency for Adversarial Robustness

Deep neural networks have been applied in many computer vision tasks and...
research
04/12/2022

Examining the Proximity of Adversarial Examples to Class Manifolds in Deep Networks

Deep neural networks achieve remarkable performance in multiple fields. ...

Please sign up or login with your details

Forgot password? Click here to reset