Measuring the Contribution of Multiple Model Representations in Detecting Adversarial Instances

11/13/2021
by   Daniel Steinberg, et al.
0

Deep learning models have been used for a wide variety of tasks. They are prevalent in computer vision, natural language processing, speech recognition, and other areas. While these models have worked well under many scenarios, it has been shown that they are vulnerable to adversarial attacks. This has led to a proliferation of research into ways that such attacks could be identified and/or defended against. Our goal is to explore the contribution that can be attributed to using multiple underlying models for the purpose of adversarial instance detection. Our paper describes two approaches that incorporate representations from multiple models for detecting adversarial examples. We devise controlled experiments for measuring the detection impact of incrementally utilizing additional models. For many of the scenarios we consider, the results show that performance increases with the number of underlying models used for extracting representations.

READ FULL TEXT
research
11/07/2022

Deviations in Representations Induced by Adversarial Attacks

Deep learning has been a popular topic and has achieved success in many ...
research
05/26/2019

Non-Determinism in Neural Networks for Adversarial Robustness

Recent breakthroughs in the field of deep learning have led to advanceme...
research
06/17/2022

Detecting Adversarial Examples in Batches – a geometrical approach

Many deep learning methods have successfully solved complex tasks in com...
research
05/18/2023

How Deep Learning Sees the World: A Survey on Adversarial Attacks Defenses

Deep Learning is currently used to perform multiple tasks, such as objec...
research
01/02/2018

Did you hear that? Adversarial Examples Against Automatic Speech Recognition

Speech is a common and effective way of communication between humans, an...
research
12/25/2018

Noise Flooding for Detecting Audio Adversarial Examples Against Automatic Speech Recognition

Neural models enjoy widespread use across a variety of tasks and have gr...
research
08/27/2023

Detecting Language Model Attacks with Perplexity

A novel hack involving Large Language Models (LLMs) has emerged, leverag...

Please sign up or login with your details

Forgot password? Click here to reset