SSSE: Efficiently Erasing Samples from Trained Machine Learning Models

07/08/2021
by   Alexandra Peşte, et al.
0

The availability of large amounts of user-provided data has been key to the success of machine learning for many real-world tasks. Recently, an increasing awareness has emerged that users should be given more control about how their data is used. In particular, users should have the right to prohibit the use of their data for training machine learning systems, and to have it erased from already trained systems. While several sample erasure methods have been proposed, all of them have drawbacks which have prevented them from gaining widespread adoption. Most methods are either only applicable to very specific families of models, sacrifice too much of the original model's accuracy, or they have prohibitive memory or computational requirements. In this paper, we propose an efficient and effective algorithm, SSSE, for samples erasure, that is applicable to a wide class of machine learning models. From a second-order analysis of the model's loss landscape we derive a closed-form update step of the model parameters that only requires access to the data to be erased, not to the original training set. Experiments on three datasets, CelebFaces attributes (CelebA), Animals with Attributes 2 (AwA2) and CIFAR10, show that in certain cases SSSE can erase samples almost as well as the optimal, yet impractical, gold standard of training a new model from scratch with only the permitted data.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
12/13/2018

Training Set Camouflage

We introduce a form of steganography in the domain of machine learning w...
research
11/01/2019

Second-Order Group Influence Functions for Black-Box Predictions

With the rapid adoption of machine learning systems in sensitive applica...
research
10/23/2018

Feasibility of Supervised Machine Learning for Cloud Security

Cloud computing is gaining significant attention, however, security is t...
research
12/09/2019

Machine Unlearning

Once users have shared their data online, it is generally difficult for ...
research
11/01/2020

Transparent Interpretation with Knockouts

How can we find a subset of training samples that are most responsible f...
research
02/03/2022

Learnability Lock: Authorized Learnability Control Through Adversarial Invertible Transformations

Owing much to the revolution of information technology, the recent progr...
research
05/13/2021

DeepObliviate: A Powerful Charm for Erasing Data Residual Memory in Deep Neural Networks

Machine unlearning has great significance in guaranteeing model security...

Please sign up or login with your details

Forgot password? Click here to reset