Remember What You Want to Forget: Algorithms for Machine Unlearning

03/04/2021
by   Ayush Sekhari, et al.
5

We study the problem of forgetting datapoints from a learnt model. In this case, the learner first receives a dataset S drawn i.i.d. from an unknown distribution, and outputs a predictor w that performs well on unseen samples from that distribution. However, at some point in the future, any training data point z ∈ S can request to be unlearned, thus prompting the learner to modify its output predictor while still ensuring the same accuracy guarantees. In our work, we initiate a rigorous study of machine unlearning in the population setting, where the goal is to maintain performance on the unseen test loss. We then provide unlearning algorithms for convex loss functions. For the setting of convex losses, we provide an unlearning algorithm that can delete up to O(n/d^1/4) samples, where d is the problem dimension. In comparison, in general, differentially private learningv(which implies unlearning) only guarantees deletion of O(n/d^1/2) samples. This shows that unlearning is at least polynomially more efficient than learning privately in terms of dependence on d in the deletion capacity.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
08/27/2019

Private Stochastic Convex Optimization with Optimal Rates

We study differentially private (DP) algorithms for stochastic convex op...
research
05/06/2022

Differentially Private Generalized Linear Models Revisited

We study the problem of (ϵ,δ)-differentially private learning of linear ...
research
02/24/2020

Approximate Data Deletion from Machine Learning Models: Algorithms and Evaluations

Deleting data from a trained machine learning (ML) model is a critical t...
research
06/15/2023

Differentially Private Domain Adaptation with Theoretical Guarantees

In many applications, the labeled data at the learner's disposal is subj...
research
05/24/2018

Learning Classifiers with Fenchel-Young Losses: Generalized Entropies, Margins, and Algorithms

We study in this paper Fenchel-Young losses, a generic way to construct ...
research
04/14/2020

Population Recovery from the Deletion Channel: Nearly Matching Trace Reconstruction Bounds

The population recovery problem asks one to recover an unknown distribut...
research
04/17/2021

Agnostic learning with unknown utilities

Traditional learning approaches for classification implicitly assume tha...

Please sign up or login with your details

Forgot password? Click here to reset