Towards Robustness against Unsuspicious Adversarial Examples

05/08/2020
by   Liang Tong, et al.
6

Despite the remarkable success of deep neural networks, significant concerns have emerged about their robustness to adversarial perturbations to inputs. While most attacks aim to ensure that these are imperceptible, physical perturbation attacks typically aim for being unsuspicious, even if perceptible. However, there is no universal notion of what it means for adversarial examples to be unsuspicious. We propose an approach for modeling suspiciousness by leveraging cognitive salience. Specifically, we split an image into foreground (salient region) and background (the rest), and allow significantly larger adversarial perturbations in the background. We describe how to compute the resulting dual-perturbation attacks on both deterministic and stochastic classifiers. We then experimentally demonstrate that our attacks do not significantly change perceptual salience of the background, but are highly effective against classifiers robust to conventional attacks. Furthermore, we show that adversarial training with dual-perturbation attacks yields classifiers that are more robust to these than state-of-the-art robust learning approaches, and comparable in terms of robustness to conventional attacks.

READ FULL TEXT

page 7

page 14

page 18

page 19

page 20

page 21

page 22

page 23

research
12/10/2018

Defending against Universal Perturbations with Shared Adversarial Training

Classifiers such as deep neural networks have been shown to be vulnerabl...
research
06/17/2021

Localized Uncertainty Attacks

The susceptibility of deep learning models to adversarial perturbations ...
research
12/15/2018

Perturbation Analysis of Learning Algorithms: A Unifying Perspective on Generation of Adversarial Examples

Despite the tremendous success of deep neural networks in various learni...
research
07/02/2020

Trace-Norm Adversarial Examples

White box adversarial perturbations are sought via iterative optimizatio...
research
12/29/2021

Repairing Adversarial Texts through Perturbation

It is known that neural networks are subject to attacks through adversar...
research
02/09/2023

Exploiting Certified Defences to Attack Randomised Smoothing

In guaranteeing that no adversarial examples exist within a bounded regi...
research
10/03/2020

Does Network Width Really Help Adversarial Robustness?

Adversarial training is currently the most powerful defense against adve...

Please sign up or login with your details

Forgot password? Click here to reset