Breaking certified defenses: Semantic adversarial examples with spoofed robustness certificates

03/19/2020
by   Amin Ghiasi, et al.
0

To deflect adversarial attacks, a range of "certified" classifiers have been proposed. In addition to labeling an image, certified classifiers produce (when possible) a certificate guaranteeing that the input image is not an ℓ_p-bounded adversarial example. We present a new attack that exploits not only the labelling function of a classifier, but also the certificate generator. The proposed method applies large perturbations that place images far from a class boundary while maintaining the imperceptibility property of adversarial examples. The proposed "Shadow Attack" causes certifiably robust networks to mislabel an image and simultaneously produce a "spoofed" certificate of robustness.

READ FULL TEXT

page 1

page 5

page 6

page 7

page 12

page 13

page 14

page 16

research
11/28/2022

Imperceptible Adversarial Attack via Invertible Neural Networks

Adding perturbations via utilizing auxiliary gradient information or dis...
research
04/17/2019

Semantic Adversarial Attacks: Parametric Transformations That Fool Deep Classifiers

Deep neural networks have been shown to exhibit an intriguing vulnerabil...
research
11/20/2019

Generate (non-software) Bugs to Fool Classifiers

In adversarial attacks intended to confound deep learning models, most s...
research
10/22/2018

Cost-Sensitive Robustness against Adversarial Examples

Several recent works have developed methods for training classifiers tha...
research
07/01/2020

Adversarial Example Games

The existence of adversarial examples capable of fooling trained neural ...
research
07/24/2021

Detecting Adversarial Examples Is (Nearly) As Hard As Classifying Them

Making classifiers robust to adversarial examples is hard. Thus, many de...
research
03/01/2021

Brain Programming is Immune to Adversarial Attacks: Towards Accurate and Robust Image Classification using Symbolic Learning

In recent years, the security concerns about the vulnerability of Deep C...

Please sign up or login with your details

Forgot password? Click here to reset