When Machine Unlearning Jeopardizes Privacy

05/05/2020
by   Min Chen, et al.
10

The right to be forgotten states that a data owner has the right to erase her data from an entity storing it. In the context of machine learning (ML), the right to be forgotten requires an ML model owner to remove the data owner's data from the training set used to build the ML model, a process known as machine unlearning. While originally designed to protect the privacy of the data owner, we argue that machine unlearning may leave some imprint of the data in the ML model and thus create unintended privacy risks. In this paper, we perform the first study on investigating the unintended information leakage caused by machine unlearning. We propose a novel membership inference attack which leverages the different outputs of an ML model's two versions to infer whether the deleted sample is part of the training set. Our experiments over five different datasets demonstrate that the proposed membership inference attack achieves strong performance. More importantly, we show that our attack in multiple cases outperforms the classical membership inference attack on the original ML model, which indicates that machine unlearning can have counterproductive effects on privacy. We notice that the privacy degradation is especially significant for well-generalized ML models where classical membership inference does not perform well. We further investigate two mechanisms to mitigate the newly discovered privacy risks and show that the only effective mechanism is to release the predicted label only. We believe that our results can help improve privacy in practical implementation of machine unlearning.

READ FULL TEXT

page 7

page 8

page 11

page 12

research
07/30/2020

Label-Leaks: Membership Inference Attack with Label

Machine learning (ML) has made tremendous progress during the past decad...
research
03/27/2021

Graph Unlearning

The right to be forgotten states that a data subject has the right to er...
research
12/09/2019

Machine Unlearning

Once users have shared their data online, it is generally difficult for ...
research
06/10/2022

Membership Inference via Backdooring

Recently issued data privacy regulations like GDPR (General Data Protect...
research
02/17/2020

Data and Model Dependencies of Membership Inference Attack

Machine Learning (ML) techniques are used by most data-driven organisati...
research
02/28/2023

Membership Inference Attack for Beluga Whales Discrimination

To efficiently monitor the growth and evolution of a particular wildlife...
research
05/15/2023

Private Training Set Inspection in MLaaS

Machine Learning as a Service (MLaaS) is a popular cloud-based solution ...

Please sign up or login with your details

Forgot password? Click here to reset