How the Softmax Output is Misleading for Evaluating the Strength of Adversarial Examples

11/21/2018
by   Utku Ozbulak, et al.
12

Even before deep learning architectures became the de facto models for complex computer vision tasks, the softmax function was, given its elegant properties, already used to analyze the predictions of feedforward neural networks. Nowadays, the output of the softmax function is also commonly used to assess the strength of adversarial examples: malicious data points designed to fail machine learning models during the testing phase. However, in this paper, we show that it is possible to generate adversarial examples that take advantage of some properties of the softmax function, leading to undesired outcomes when interpreting the strength of the adversarial examples at hand. Specifically, we argue that the output of the softmax function is a poor indicator when the strength of an adversarial example is analyzed and that this indicator can be easily tricked by already existing methods for adversarial example generation.

READ FULL TEXT

page 2

page 3

research
07/30/2019

Not All Adversarial Examples Require a Complex Defense: Identifying Over-optimized Adversarial Examples with IQR-based Logit Thresholding

Detecting adversarial examples currently stands as one of the biggest ch...
research
02/22/2018

Adversarial Examples that Fool both Human and Computer Vision

Machine learning models are vulnerable to adversarial examples: small ch...
research
01/22/2020

Elephant in the Room: An Evaluation Framework for Assessing Adversarial Examples in NLP

An adversarial example is an input transformed by small perturbations th...
research
11/22/2018

Distorting Neural Representations to Generate Highly Transferable Adversarial Examples

Deep neural networks (DNN) can be easily fooled by adding human impercep...
research
06/21/2021

Adversarial Examples Make Strong Poisons

The adversarial machine learning literature is largely partitioned into ...
research
04/08/2018

Adaptive Spatial Steganography Based on Probability-Controlled Adversarial Examples

Deep learning model is vulnerable to adversarial attack, which generates...
research
03/09/2018

On Generation of Adversarial Examples using Convex Programming

It has been observed that deep learning architectures tend to make erron...

Please sign up or login with your details

Forgot password? Click here to reset