Defending against Adversarial Attacks through Resilient Feature Regeneration

06/08/2019
by   Tejas Borkar, et al.
6

Deep neural network (DNN) predictions have been shown to be vulnerable to carefully crafted adversarial perturbations. Specifically, so-called universal adversarial perturbations are image-agnostic perturbations that can be added to any image and can fool a target network into making erroneous predictions. Departing from existing adversarial defense strategies, which work in the image domain, we present a novel defense which operates in the DNN feature domain and effectively defends against such universal adversarial attacks. Our approach identifies pre-trained convolutional features that are most vulnerable to adversarial noise and deploys defender units which transform (regenerate) these DNN filter activations into noise-resilient features, guarding against unseen adversarial perturbations. The proposed defender units are trained using a target loss on synthetic adversarial perturbations, which we generate with a novel efficient synthesis method. We validate the proposed method for different DNN architectures, and demonstrate that it outperforms existing defense strategies across network architectures by more than 10 Moreover, we demonstrate that the approach also improves resilience of DNNs to other unseen adversarial attacks.

READ FULL TEXT

page 1

page 5

page 7

page 8

page 12

page 14

research
11/12/2019

Robust Design of Deep Neural Networks against Adversarial Attacks based on Lyapunov Theory

Deep neural networks (DNNs) are vulnerable to subtle adversarial perturb...
research
07/13/2022

Perturbation Inactivation Based Adversarial Defense for Face Recognition

Deep learning-based face recognition models are vulnerable to adversaria...
research
11/02/2019

Security of Facial Forensics Models Against Adversarial Attacks

Deep neural networks (DNNs) have been used in forensics to identify fake...
research
08/12/2019

Adversarial Neural Pruning

It is well known that neural networks are susceptible to adversarial per...
research
03/03/2019

A Kernelized Manifold Mapping to Diminish the Effect of Adversarial Perturbations

The linear and non-flexible nature of deep convolutional models makes th...
research
05/25/2023

Don't Retrain, Just Rewrite: Countering Adversarial Perturbations by Rewriting Text

Can language models transform inputs to protect text classifiers against...
research
02/20/2022

Real-time Over-the-air Adversarial Perturbations for Digital Communications using Deep Neural Networks

Deep neural networks (DNNs) are increasingly being used in a variety of ...

Please sign up or login with your details

Forgot password? Click here to reset