DP-InstaHide: Provably Defusing Poisoning and Backdoor Attacks with Differentially Private Data Augmentations

03/02/2021
by   Eitan Borgnia, et al.
0

Data poisoning and backdoor attacks manipulate training data to induce security breaches in a victim model. These attacks can be provably deflected using differentially private (DP) training methods, although this comes with a sharp decrease in model performance. The InstaHide method has recently been proposed as an alternative to DP training that leverages supposed privacy properties of the mixup augmentation, although without rigorous guarantees. In this work, we show that strong data augmentations, such as mixup and random additive noise, nullify poison attacks while enduring only a small accuracy trade-off. To explain these finding, we propose a training method, DP-InstaHide, which combines the mixup regularizer with additive noise. A rigorous analysis of DP-InstaHide shows that mixup does indeed have privacy advantages, and that training with k-way mixup provably yields at least k times stronger DP guarantees than a naive DP mechanism. Because mixup (as opposed to noise) is beneficial to model performance, DP-InstaHide provides a mechanism for achieving stronger empirical performance against poisoning attacks than other known DP methods.

READ FULL TEXT

page 4

page 7

research
05/02/2023

Differentially Private In-Context Learning

An important question in deploying large language models (LLMs) is how t...
research
10/02/2019

Improving Differentially Private Models with Active Learning

Broad adoption of machine learning techniques has increased privacy conc...
research
11/18/2020

Strong Data Augmentation Sanitizes Poisoning and Backdoor Attacks Without an Accuracy Tradeoff

Data poisoning and backdoor attacks manipulate victim models by maliciou...
research
02/20/2020

Differentially Private ERM Based on Data Perturbation

In this paper, after observing that different training data instances af...
research
07/01/2023

Saibot: A Differentially Private Data Search Platform

Recent data search platforms use ML task-based utility measures rather t...
research
07/24/2023

A Differentially Private Weighted Empirical Risk Minimization Procedure and its Application to Outcome Weighted Learning

It is commonplace to use data containing personal information to build p...
research
04/21/2022

Differentially Private Learning with Margin Guarantees

We present a series of new differentially private (DP) algorithms with d...

Please sign up or login with your details

Forgot password? Click here to reset