DiffDefense: Defending against Adversarial Attacks via Diffusion Models

This paper presents a novel reconstruction method that leverages Diffusion Models to protect machine learning classifiers against adversarial attacks, all without requiring any modifications to the classifiers themselves. The susceptibility of machine learning models to minor input perturbations renders them vulnerable to adversarial attacks. While diffusion-based methods are typically disregarded for adversarial defense due to their slow reverse process, this paper demonstrates that our proposed method offers robustness against adversarial threats while preserving clean accuracy, speed, and plug-and-play compatibility. Code at: https://github.com/HondamunigePrasannaSilva/DiffDefence.

READ FULL TEXT
research
09/24/2020

Torchattacks : A Pytorch Repository for Adversarial Attacks

Torchattacks is a PyTorch library that contains adversarial attacks to g...
research
01/17/2023

Denoising Diffusion Probabilistic Models as a Defense against Adversarial Attacks

Neural Networks are infamously sensitive to small perturbations in their...
research
05/30/2022

Guided Diffusion Model for Adversarial Purification

With wider application of deep neural networks (DNNs) in various algorit...
research
02/11/2022

Using Random Perturbations to Mitigate Adversarial Attacks on Sentiment Analysis Models

Attacks on deep learning models are often difficult to identify and ther...
research
03/17/2023

Adversarial Counterfactual Visual Explanations

Counterfactual explanations and adversarial attacks have a related goal:...
research
10/22/2022

Hindering Adversarial Attacks with Implicit Neural Representations

We introduce the Lossy Implicit Network Activation Coding (LINAC) defenc...
research
03/02/2023

Defending against Adversarial Audio via Diffusion Model

Deep learning models have been widely used in commercial acoustic system...

Please sign up or login with your details

Forgot password? Click here to reset