Guided Diffusion Model for Adversarial Purification

05/30/2022
by   Jinyi Wang, et al.
3

With wider application of deep neural networks (DNNs) in various algorithms and frameworks, security threats have become one of the concerns. Adversarial attacks disturb DNN-based image classifiers, in which attackers can intentionally add imperceptible adversarial perturbations on input images to fool the classifiers. In this paper, we propose a novel purification approach, referred to as guided diffusion model for purification (GDMP), to help protect classifiers from adversarial attacks. The core of our approach is to embed purification into the diffusion denoising process of a Denoised Diffusion Probabilistic Model (DDPM), so that its diffusion process could submerge the adversarial perturbations with gradually added Gaussian noises, and both of these noises can be simultaneously removed following a guided denoising process. On our comprehensive experiments across various datasets, the proposed GDMP is shown to reduce the perturbations raised by adversarial attacks to a shallow range, thereby significantly improving the correctness of classification. GDMP improves the robust accuracy by 5 PGD attack on the CIFAR10 dataset. Moreover, GDMP achieves 70.94 the challenging ImageNet dataset.

READ FULL TEXT

page 2

page 5

page 6

research
01/17/2023

Denoising Diffusion Probabilistic Models as a Defense against Adversarial Attacks

Neural Networks are infamously sensitive to small perturbations in their...
research
05/01/2022

DDDM: a Brain-Inspired Framework for Robust Classification

Despite their outstanding performance in a broad spectrum of real-world ...
research
03/26/2022

Reverse Engineering of Imperceptible Adversarial Image Perturbations

It has been well recognized that neural network based image classifiers ...
research
06/11/2021

Adversarial purification with Score-based generative models

While adversarial training is considered as a standard defense method ag...
research
09/07/2023

DiffDefense: Defending against Adversarial Attacks via Diffusion Models

This paper presents a novel reconstruction method that leverages Diffusi...
research
06/22/2022

Guided Diffusion Model for Adversarial Purification from Random Noise

In this paper, we propose a novel guided diffusion purification approach...
research
11/02/2019

Security of Facial Forensics Models Against Adversarial Attacks

Deep neural networks (DNNs) have been used in forensics to identify fake...

Please sign up or login with your details

Forgot password? Click here to reset