Counterstrike: Defending Deep Learning Architectures Against Adversarial Samples by Langevin Dynamics with Supervised Denoising Autoencoder

05/30/2018
by   Vignesh Srinivasan, et al.
2

Adversarial attacks on deep learning models have been demonstrated to be imperceptible to a human, while decreasing the model performance considerably. Attempts to provide invariance against such attacks have denoised adversarial samples to only send cleaned samples to the classifier. In a similar spirit this paper proposes a novel effective strategy that allows to relax adversarial samples onto the underlying manifold of the (unknown) target class distribution. Specifically, given an off-manifold adversarial example, our Metroplis-adjusted Langevin algorithm (Mala) guided through a supervised denoising autoencoder network (sDAE) allows to drive the adversarial samples towards high density regions of the data generating distribution. So, in a nutshell the adversarial example is transformed back from off-manifold onto the data manifold for which the learning model was originally trained and where it can perform well and robustly. Experiments on various benchmark datasets show that our novel Malade method exhibits a high robustness against blackbox and whitebox attacks and outperforms state-of-the-art defense algorithms.

READ FULL TEXT

page 2

page 3

page 14

page 15

page 16

page 17

page 18

research
09/05/2020

Dual Manifold Adversarial Robustness: Defense against Lp and non-Lp Adversarial Attacks

Adversarial training is a popular defense strategy against attack threat...
research
03/28/2023

Denoising Autoencoder-based Defensive Distillation as an Adversarial Robustness Algorithm

Adversarial attacks significantly threaten the robustness of deep neural...
research
01/30/2023

Identifying Adversarially Attackable and Robust Samples

This work proposes a novel perspective on adversarial attacks by introdu...
research
06/13/2019

A Computationally Efficient Method for Defending Adversarial Deep Learning Attacks

The reliance on deep learning algorithms has grown significantly in rece...
research
12/07/2018

Combatting Adversarial Attacks through Denoising and Dimensionality Reduction: A Cascaded Autoencoder Approach

Machine Learning models are vulnerable to adversarial attacks that rely ...
research
05/27/2018

Defending Against Adversarial Attacks by Leveraging an Entire GAN

Recent work has shown that state-of-the-art models are highly vulnerable...
research
08/02/2022

Success of Uncertainty-Aware Deep Models Depends on Data Manifold Geometry

For responsible decision making in safety-critical settings, machine lea...

Please sign up or login with your details

Forgot password? Click here to reset