Adversarial Pixel Restoration as a Pretext Task for Transferable Perturbations

07/18/2022
by   Hashmat Shadab Malik, et al.
3

Transferable adversarial attacks optimize adversaries from a pretrained surrogate model and known label space to fool the unknown black-box models. Therefore, these attacks are restricted by the availability of an effective surrogate model. In this work, we relax this assumption and propose Adversarial Pixel Restoration as a self-supervised alternative to train an effective surrogate model from scratch under the condition of no labels and few data samples. Our training approach is based on a min-max objective which reduces overfitting via an adversarial objective and thus optimizes for a more generalizable surrogate model. Our proposed attack is complimentary to our adversarial pixel restoration and is independent of any task specific objective as it can be launched in a self-supervised manner. We successfully demonstrate the adversarial transferability of our approach to Vision Transformers as well as Convolutional Neural Networks for the tasks of classification, object detection, and video segmentation. Our codes pre-trained surrogate models are available at: https://github.com/HashmatShadab/APR

READ FULL TEXT

page 1

page 3

page 6

page 9

page 16

research
02/23/2023

Boosting Adversarial Transferability using Dynamic Cues

The transferability of adversarial perturbations between image models ha...
research
10/12/2022

Boosting the Transferability of Adversarial Attacks with Reverse Adversarial Perturbation

Deep neural networks (DNNs) have been shown to be vulnerable to adversar...
research
08/24/2023

Exploring Transferability of Multimodal Adversarial Samples for Vision-Language Pre-training Models with Contrastive Learning

Vision-language pre-training models (VLP) are vulnerable, especially to ...
research
06/08/2020

A Self-supervised Approach for Adversarial Robustness

Adversarial examples can cause catastrophic mistakes in Deep Neural Netw...
research
11/03/2022

Data-free Defense of Black Box Models Against Adversarial Attacks

Several companies often safeguard their trained deep models (i.e. detail...
research
01/18/2022

Dilated Convolutional Neural Networks for Lightweight Diacritics Restoration

Diacritics restoration has become a ubiquitous task in the Latin-alphabe...
research
07/25/2023

On the unreasonable vulnerability of transformers for image restoration – and an easy fix

Following their success in visual recognition tasks, Vision Transformers...

Please sign up or login with your details

Forgot password? Click here to reset