Rethinking the Trigger of Backdoor Attack

04/09/2020
by   Yiming Li, et al.
7

In this work, we study the problem of backdoor attacks, which add a specific trigger (i.e., a local patch) onto some training images to enforce that the testing images with the same trigger are incorrectly predicted while the natural testing examples are correctly predicted by the trained model. Many existing works adopted the setting that the triggers across the training and testing images follow the same appearance and are located at the same area. However, we observe that if the appearance or location of the trigger is slightly changed, then the attack performance may degrade sharply. According to this observation, we propose to spatially transform (e.g., flipping and scaling) the testing image, such that the appearance and location of the trigger (if exists) will be changed. This simple strategy is experimentally verified to be effective to defend many state-of-the-art backdoor attack methods. Furthermore, to enhance the robustness of the backdoor attacks, we propose to conduct the random spatial transformation on the training images with the trigger before feeding into the training process. Extensive experiments verify that the proposed backdoor attack is robust to spatial transformations.

READ FULL TEXT

page 4

page 6

page 8

page 11

page 12

research
04/06/2021

Backdoor Attack in the Physical World

Backdoor attack intends to inject hidden backdoor into the deep neural n...
research
11/02/2022

BATT: Backdoor Attack with Transformation-based Triggers

Deep neural networks (DNNs) are vulnerable to backdoor attacks. The back...
research
07/24/2019

Joint Adversarial Training: Incorporating both Spatial and Pixel Attacks

Conventional adversarial training methods using attacks that manipulate ...
research
05/01/2020

Jacks of All Trades, Masters Of None: Addressing Distributional Shift and Obtrusiveness via Transparent Patch Attacks

We focus on the development of effective adversarial patch attacks and –...
research
05/24/2023

Sharpness-Aware Data Poisoning Attack

Recent research has highlighted the vulnerability of Deep Neural Network...
research
06/30/2021

Local Reweighting for Adversarial Training

Instances-reweighted adversarial training (IRAT) can significantly boost...

Please sign up or login with your details

Forgot password? Click here to reset