Robustness Out of the Box: Compositional Representations Naturally Defend Against Black-Box Patch Attacks

12/01/2020
by   Christian Cosgrove, et al.
0

Patch-based adversarial attacks introduce a perceptible but localized change to the input that induces misclassification. While progress has been made in defending against imperceptible attacks, it remains unclear how patch-based attacks can be resisted. In this work, we study two different approaches for defending against black-box patch attacks. First, we show that adversarial training, which is successful against imperceptible attacks, has limited effectiveness against state-of-the-art location-optimized patch attacks. Second, we find that compositional deep networks, which have part-based representations that lead to innate robustness to natural occlusion, are robust to patch attacks on PASCAL3D+ and the German Traffic Sign Recognition Benchmark, without adversarial training. Moreover, the robustness of compositional models outperforms that of adversarially trained standard models by a large margin. However, on GTSRB, we observe that they have problems discriminating between similar traffic signs with fine-grained differences. We overcome this limitation by introducing part-based finetuning, which improves fine-grained recognition. By leveraging compositional representations, this is the first work that defends against black-box patch attacks without expensive adversarial training. This defense is more robust than adversarial training and more interpretable because it can locate and ignore adversarial patches.

READ FULL TEXT

page 1

page 4

page 7

page 8

research
07/24/2021

Adversarial training may be a double-edged sword

Adversarial training has been shown as an effective approach to improve ...
research
06/28/2023

Distributional Modeling for Location-Aware Adversarial Patches

Adversarial patch is one of the important forms of performing adversaria...
research
04/12/2020

PatchAttack: A Black-box Texture-based Attack with Reinforcement Learning

Patch-based attacks introduce a perceptible but localized change to the ...
research
06/10/2020

Towards Robust Fine-grained Recognition by Maximal Separation of Discriminative Features

Adversarial attacks have been widely studied for general classification ...
research
02/07/2021

SPADE: A Spectral Method for Black-Box Adversarial Robustness Evaluation

A black-box spectral method is introduced for evaluating the adversarial...
research
12/05/2021

Stochastic Local Winner-Takes-All Networks Enable Profound Adversarial Robustness

This work explores the potency of stochastic competition-based activatio...
research
03/21/2023

Efficient Decision-based Black-box Patch Attacks on Video Recognition

Although Deep Neural Networks (DNNs) have demonstrated excellent perform...

Please sign up or login with your details

Forgot password? Click here to reset