Locally optimal detection of stochastic targeted universal adversarial perturbations

12/08/2020
by   Amish Goel, et al.
0

Deep learning image classifiers are known to be vulnerable to small adversarial perturbations of input images. In this paper, we derive the locally optimal generalized likelihood ratio test (LO-GLRT) based detector for detecting stochastic targeted universal adversarial perturbations (UAPs) of the classifier inputs. We also describe a supervised training method to learn the detector's parameters, and demonstrate better performance of the detector compared to other detection methods on several popular image classification datasets.

READ FULL TEXT
research
10/10/2022

Universal Adversarial Perturbations: Efficiency on a small image dataset

Although neural networks perform very well on the image classification t...
research
12/31/2021

On Distinctive Properties of Universal Perturbations

We identify properties of universal adversarial perturbations (UAPs) tha...
research
11/16/2017

Defense against Universal Adversarial Perturbations

Recent advances in Deep Learning show the existence of image-agnostic qu...
research
02/14/2017

On Detecting Adversarial Perturbations

Machine learning and deep learning in particular has advanced tremendous...
research
08/08/2019

Universal Adversarial Audio Perturbations

We demonstrate the existence of universal adversarial perturbations, whi...
research
10/15/2020

Adversarial Images through Stega Glasses

This paper explores the connection between steganography and adversarial...
research
08/01/2016

Early Methods for Detecting Adversarial Images

Many machine learning classifiers are vulnerable to adversarial perturba...

Please sign up or login with your details

Forgot password? Click here to reset