Detecting Adversarial Examples in Batches – a geometrical approach

06/17/2022
by   Danush Kumar Venkatesh, et al.
11

Many deep learning methods have successfully solved complex tasks in computer vision and speech recognition applications. Nonetheless, the robustness of these models has been found to be vulnerable to perturbed inputs or adversarial examples, which are imperceptible to the human eye, but lead the model to erroneous output decisions. In this study, we adapt and introduce two geometric metrics, density and coverage, and evaluate their use in detecting adversarial samples in batches of unseen data. We empirically study these metrics using MNIST and two real-world biomedical datasets from MedMNIST, subjected to two different adversarial attacks. Our experiments show promising results for both metrics to detect adversarial examples. We believe that his work can lay the ground for further study on these metrics' use in deployed machine learning systems to monitor for possible attacks by adversarial examples or related pathologies such as dataset shift.

READ FULL TEXT

page 3

page 11

page 12

page 13

page 14

page 15

page 16

page 17

research
11/08/2019

Imperceptible Adversarial Attacks on Tabular Data

Security of machine learning models is a concern as they may face advers...
research
09/11/2018

Isolated and Ensemble Audio Preprocessing Methods for Detecting Adversarial Examples against Automatic Speech Recognition

An adversarial attack is an exploitative process in which minute alterat...
research
11/24/2019

Robustness Metrics for Real-World Adversarial Examples

We explore metrics to evaluate the robustness of real-world adversarial ...
research
01/28/2021

Increasing the Confidence of Deep Neural Networks by Coverage Analysis

The great performance of machine learning algorithms and deep neural net...
research
04/30/2019

Detecting Adversarial Examples through Nonlinear Dimensionality Reduction

Deep neural networks are vulnerable to adversarial examples, i.e., caref...
research
06/01/2020

Adversarial Attacks on Reinforcement Learning based Energy Management Systems of Extended Range Electric Delivery Vehicles

Adversarial examples are firstly investigated in the area of computer vi...
research
11/13/2021

Measuring the Contribution of Multiple Model Representations in Detecting Adversarial Instances

Deep learning models have been used for a wide variety of tasks. They ar...

Please sign up or login with your details

Forgot password? Click here to reset