Invisible Backdoor Attacks Against Deep Neural Networks

09/06/2019
by   Shaofeng Li, et al.
0

Deep neural networks (DNNs) have been proven vulnerable to backdoor attacks, where hidden features (patterns) trained to a normal model, and only activated by some specific input (called triggers), trick the model into producing unexpected behavior. In this paper, we design an optimization framework to create covert and scattered triggers for backdoor attacks, invisible backdoors, where triggers can amplify the specific neuron activation, while being invisible to both backdoor detection methods and human inspection. We use the Perceptual Adversarial Similarity Score (PASS) rozsa2016adversarial to define invisibility for human users and apply L_2 and L_0 regularization into the optimization process to hide the trigger within the input data. We show that the proposed invisible backdoors can be fairly effective across various DNN models as well as three datasets CIFAR-10, CIFAR-100, and GTSRB, by measuring their attack success rates and invisibility scores.

READ FULL TEXT

page 3

page 6

page 7

research
02/26/2020

Defending against Backdoor Attack on Deep Neural Networks

Although deep neural networks (DNNs) have achieved a great success in va...
research
05/10/2023

Towards Invisible Backdoor Attacks in the Frequency Domain against Deep Neural Networks

Deep neural networks (DNNs) have made tremendous progress in the past te...
research
05/26/2022

BppAttack: Stealthy and Efficient Trojan Attacks against Deep Neural Networks via Image Quantization and Contrastive Adversarial Learning

Deep neural networks are vulnerable to Trojan attacks. Existing attacks ...
research
03/17/2022

PiDAn: A Coherence Optimization Approach for Backdoor Attack Detection and Mitigation in Deep Neural Networks

Backdoor attacks impose a new threat in Deep Neural Networks (DNNs), whe...
research
04/27/2022

Detecting Backdoor Poisoning Attacks on Deep Neural Networks by Heatmap Clustering

Predicitions made by neural networks can be fraudulently altered by so-c...
research
10/10/2019

Coloring the Black Box: Visualizing neural network behavior with a self-introspective model

The following work presents how autoencoding all the possible hidden act...
research
05/01/2021

Hidden Backdoors in Human-Centric Language Models

Natural language processing (NLP) systems have been proven to be vulnera...

Please sign up or login with your details

Forgot password? Click here to reset