Detecting Adversarial Examples through Nonlinear Dimensionality Reduction

04/30/2019
by   Francesco Crecchi, et al.
16

Deep neural networks are vulnerable to adversarial examples, i.e., carefully-perturbed inputs aimed to mislead classification. This work proposes a detection method based on combining non-linear dimensionality reduction and density estimation techniques. Our empirical findings show that the proposed approach is able to effectively detect adversarial examples crafted by non-adaptive attackers, i.e., not specifically tuned to bypass the detection method. Given our promising results, we plan to extend our analysis to adaptive attackers in future work.

READ FULL TEXT
research
03/09/2018

Detecting Adversarial Examples - A Lesson from Multimedia Forensics

Adversarial classification is the task of performing robust classificati...
research
05/20/2017

Adversarial Examples Are Not Easily Detected: Bypassing Ten Detection Methods

Neural networks are known to be vulnerable to adversarial examples: inpu...
research
06/17/2022

Detecting Adversarial Examples in Batches – a geometrical approach

Many deep learning methods have successfully solved complex tasks in com...
research
08/23/2017

Is Deep Learning Safe for Robot Vision? Adversarial Examples against the iCub Humanoid

Deep neural networks have been widely adopted in recent years, exhibitin...
research
10/01/2019

Deep Neural Rejection against Adversarial Examples

Despite the impressive performances reported by deep neural networks in ...
research
01/01/2022

Adversarial Attack via Dual-Stage Network Erosion

Deep neural networks are vulnerable to adversarial examples, which can f...
research
04/17/2019

Interpreting Adversarial Examples with Attributes

Deep computer vision systems being vulnerable to imperceptible and caref...

Please sign up or login with your details

Forgot password? Click here to reset