Imperceptible Backdoor Attack: From Input Space to Feature Representation

05/06/2022
by   Nan Zhong, et al.
0

Backdoor attacks are rapidly emerging threats to deep neural networks (DNNs). In the backdoor attack scenario, attackers usually implant the backdoor into the target model by manipulating the training dataset or training process. Then, the compromised model behaves normally for benign input yet makes mistakes when the pre-defined trigger appears. In this paper, we analyze the drawbacks of existing attack approaches and propose a novel imperceptible backdoor attack. We treat the trigger pattern as a special kind of noise following a multinomial distribution. A U-net-based network is employed to generate concrete parameters of multinomial distribution for each benign input. This elaborated trigger ensures that our approach is invisible to both humans and statistical detection. Besides the design of the trigger, we also consider the robustness of our approach against model diagnose-based defences. We force the feature representation of malicious input stamped with the trigger to be entangled with the benign one. We demonstrate the effectiveness and robustness against multiple state-of-the-art defences through extensive datasets and networks. Our trigger only modifies less than 1% pixels of a benign image while the modification magnitude is 1. Our source code is available at https://github.com/Ekko-zn/IJCAI2022-Backdoor.

READ FULL TEXT

page 3

page 5

research
06/15/2020

An Embarrassingly Simple Approach for Trojan Attack in Deep Neural Networks

With the widespread use of deep neural networks (DNNs) in high-stake app...
research
10/20/2021

Moiré Attack (MA): A New Potential Risk of Screen Photos

Images, captured by a camera, play a critical role in training Deep Neur...
research
07/27/2022

Hardly Perceptible Trojan Attack against Neural Networks with Bit Flips

The security of deep neural networks (DNNs) has attracted increasing att...
research
09/10/2019

FDA: Feature Disruptive Attack

Though Deep Neural Networks (DNN) show excellent performance across vari...
research
07/16/2020

Deep Learning Backdoors

Intuitively, a backdoor attack against Deep Neural Networks (DNNs) is to...
research
09/18/2020

The Hidden Vulnerability of Watermarking for Deep Neural Networks

Watermarking has shown its effectiveness in protecting the intellectual ...
research
11/18/2019

NeuronInspect: Detecting Backdoors in Neural Networks via Output Explanations

Deep neural networks have achieved state-of-the-art performance on vario...

Please sign up or login with your details

Forgot password? Click here to reset