Hard to Forget: Poisoning Attacks on Certified Machine Unlearning

09/17/2021
by   Neil G. Marchant, et al.
0

The right to erasure requires removal of a user's information from data held by organizations, with rigorous interpretations extending to downstream products such as learned models. Retraining from scratch with the particular user's data omitted fully removes its influence on the resulting model, but comes with a high computational cost. Machine "unlearning" mitigates the cost incurred by full retraining: instead, models are updated incrementally, possibly only requiring retraining when approximation errors accumulate. Rapid progress has been made towards privacy guarantees on the indistinguishability of unlearned and retrained models, but current formalisms do not place practical bounds on computation. In this paper we demonstrate how an attacker can exploit this oversight, highlighting a novel attack surface introduced by machine unlearning. We consider an attacker aiming to increase the computational cost of data removal. We derive and empirically investigate a poisoning attack on certified machine unlearning where strategically designed training data triggers complete retraining when removed.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
11/08/2019

Certified Data Removal from Machine Learning Models

Good data stewardship requires removal of data at the request of the dat...
research
07/08/2020

Dung's semantics satisfy attack removal monotonicity

We show that preferred, stable, complete, and grounded semantics satisfy...
research
11/15/2019

Resource-Competitive Sybil Defenses

Proof-of-work(PoW) is an algorithmic tool used to secure networks by imp...
research
12/23/2019

Privacy Attacks on Network Embeddings

Data ownership and data protection are increasingly important topics wit...
research
06/29/2022

Approximate Data Deletion in Generative Models

Users have the right to have their data deleted by third-party learned s...
research
09/02/2022

An Introduction to Machine Unlearning

Removing the influence of a specified subset of training data from a mac...

Please sign up or login with your details

Forgot password? Click here to reset