Class-Oriented Poisoning Attack

07/31/2020
by   Bingyin Zhao, et al.
10

Poisoning attacks on machine learning systems compromise the model performance by deliberately injecting malicious samples in the training dataset to influence the training process. Prior works focus on either availability attacks (i.e., lowering the overall model accuracy) or integrity attacks (i.e., enabling specific instance based backdoor). In this paper, we advance the adversarial objectives of the availability attacks to a per-class basis, which we refer to as class-oriented poisoning attacks. We demonstrate that the proposed attack is capable of forcing the corrupted model to predict in two specific ways: (i) classify unseen new images to a targeted "supplanter" class, and (ii) misclassify images from a "victim" class while maintaining the classification accuracy on other non-victim classes. To maximize the adversarial effect, we propose a gradient-based framework that manipulates the logits to retain/eliminate the desired/undesired feature information in the generated poisoning images. Using newly defined metrics at the class level, we illustrate the effectiveness of the proposed class-oriented poisoning attacks on various models (e.g., LeNet-5, Vgg-9, and ResNet-50) over a wide range of datasets (e.g., MNIST, CIFAR-10, and ImageNet-ILSVRC2012).

READ FULL TEXT

page 1

page 2

page 4

page 5

page 6

page 7

page 9

page 10

research
02/11/2021

Adversarial Poisoning Attacks and Defense for General Multi-Class Models Based On Synthetic Reduced Nearest Neighbors

State-of-the-art machine learning models are vulnerable to data poisonin...
research
06/08/2020

Tricking Adversarial Attacks To Fail

Recent adversarial defense approaches have failed. Untargeted gradient-b...
research
04/19/2021

Manipulating SGD with Data Ordering Attacks

Machine learning is vulnerable to a wide variety of different attacks. I...
research
07/24/2020

T-BFA: Targeted Bit-Flip Adversarial Weight Attack

Deep Neural Network (DNN) attacks have mostly been conducted through adv...
research
05/09/2022

Btech thesis report on adversarial attack detection and purification of adverserially attacked images

This is Btech thesis report on detection and purification of adverserial...
research
10/20/2022

New data poison attacks on machine learning classifiers for mobile exfiltration

Most recent studies have shown several vulnerabilities to attacks with t...
research
08/17/2018

Reinforcement Learning for Autonomous Defence in Software-Defined Networking

Despite the successful application of machine learning (ML) in a wide ra...

Please sign up or login with your details

Forgot password? Click here to reset