Reverse Engineering of Imperceptible Adversarial Image Perturbations

03/26/2022
by   Yifan Gong, et al.
4

It has been well recognized that neural network based image classifiers are easily fooled by images with tiny perturbations crafted by an adversary. There has been a vast volume of research to generate and defend such adversarial attacks. However, the following problem is left unexplored: How to reverse-engineer adversarial perturbations from an adversarial image? This leads to a new adversarial learning paradigm–Reverse Engineering of Deceptions (RED). If successful, RED allows us to estimate adversarial perturbations and recover the original images. However, carefully crafted, tiny adversarial perturbations are difficult to recover by optimizing a unilateral RED objective. For example, the pure image denoising method may overfit to minimizing the reconstruction error but hardly preserve the classification properties of the true adversarial perturbations. To tackle this challenge, we formalize the RED problem and identify a set of principles crucial to the RED approach design. Particularly, we find that prediction alignment and proper data augmentation (in terms of spatial transformations) are two criteria to achieve a generalizable RED approach. By integrating these RED principles with image denoising, we propose a new Class-Discriminative Denoising based RED framework, termed CDD-RED. Extensive experiments demonstrate the effectiveness of CDD-RED under different evaluation metrics (ranging from the pixel-level, prediction-level to the attribution-level alignment) and a variety of attack generation methods (e.g., FGSM, PGD, CW, AutoAttack, and adaptive attacks).

READ FULL TEXT

page 4

page 7

page 19

research
05/30/2022

Guided Diffusion Model for Adversarial Purification

With wider application of deep neural networks (DNNs) in various algorit...
research
01/26/2018

Deflecting Adversarial Attacks with Pixel Deflection

CNNs are poised to become integral parts of many critical systems. Despi...
research
12/09/2017

NAG: Network for Adversary Generation

Adversarial perturbations can pose a serious threat for deploying machin...
research
11/02/2019

Security of Facial Forensics Models Against Adversarial Attacks

Deep neural networks (DNNs) have been used in forensics to identify fake...
research
08/27/2018

Targeted Nonlinear Adversarial Perturbations in Images and Videos

We introduce a method for learning adversarial perturbations targeted to...
research
03/09/2022

Reverse Engineering ℓ_p attacks: A block-sparse optimization approach with recovery guarantees

Deep neural network-based classifiers have been shown to be vulnerable t...
research
09/12/2019

A method for Cloud Mapping in the Field of View of the Infra-Red Camera during the EUSO-SPB1 flight

EUSO-SPB1 was released on April 24th, 2017, from the NASA balloon launch...

Please sign up or login with your details

Forgot password? Click here to reset