Hidden Trigger Backdoor Attacks

09/30/2019
by   Aniruddha Saha, et al.
0

With the success of deep learning algorithms in various domains, studying adversarial attacks to secure deep models in real world applications has become an important research topic. Backdoor attacks are a form of adversarial attacks on deep networks where the attacker provides poisoned data to the victim to train the model with, and then activates the attack by showing a specific trigger pattern at the test time. Most state-of-the-art backdoor attacks either provide mislabeled poisoning data that is possible to identify by visual inspection, reveal the trigger in the poisoned data, or use noise and perturbation to hide the trigger. We propose a novel form of backdoor attack where poisoned data look natural with correct labels and also more importantly, the attacker hides the trigger in the poisoned data and keeps the trigger secret until the test time. We perform an extensive study on various image classification settings and show that our attack can fool the model by pasting the trigger at random locations on unseen images although the model performs well on clean data. We also show that our proposed attack cannot be easily defended using a state-of-the-art defense algorithm for backdoor attacks.

READ FULL TEXT

page 2

page 6

research
11/05/2020

Defense-friendly Images in Adversarial Attacks: Dataset and Metrics for Perturbation Difficulty

Dataset bias is a problem in adversarial machine learning, especially in...
research
02/20/2021

WaNet – Imperceptible Warping-based Backdoor Attack

With the thriving of deep learning and the widespread practice of using ...
research
10/19/2021

Multi-concept adversarial attacks

As machine learning (ML) techniques are being increasingly used in many ...
research
09/23/2021

Adversarial Transfer Attacks With Unknown Data and Class Overlap

The ability to transfer adversarial attacks from one model (the surrogat...
research
09/02/2021

Excess Capacity and Backdoor Poisoning

A backdoor data poisoning attack is an adversarial attack wherein the at...
research
07/26/2020

Robust Collective Classification against Structural Attacks

Collective learning methods exploit relations among data points to enhan...
research
07/21/2020

Backdoor Attacks and Countermeasures on Deep Learning: A Comprehensive Review

This work provides the community with a timely comprehensive review of b...

Please sign up or login with your details

Forgot password? Click here to reset