DeepObliviate: A Powerful Charm for Erasing Data Residual Memory in Deep Neural Networks

05/13/2021
by   Yingzhe He, et al.
0

Machine unlearning has great significance in guaranteeing model security and protecting user privacy. Additionally, many legal provisions clearly stipulate that users have the right to demand model providers to delete their own data from training set, that is, the right to be forgotten. The naive way of unlearning data is to retrain the model without it from scratch, which becomes extremely time and resource consuming at the modern scale of deep neural networks. Other unlearning approaches by refactoring model or training data struggle to gain a balance between overhead and model usability. In this paper, we propose an approach, dubbed as DeepObliviate, to implement machine unlearning efficiently, without modifying the normal training mode. Our approach improves the original training process by storing intermediate models on the hard disk. Given a data point to unlearn, we first quantify its temporal residual memory left in stored models. The influenced models will be retrained and we decide when to terminate the retraining based on the trend of residual memory on-the-fly. Last, we stitch an unlearned model by combining the retrained models and uninfluenced models. We extensively evaluate our approach on five datasets and deep learning models. Compared to the method of retraining from scratch, our approach can achieve 99.0 accuracy rates and 66.7×, 75.0×, 33.3×, 29.4×, 13.7× speedups on the MNIST, SVHN, CIFAR-10, Purchase, and ImageNet datasets, respectively. Compared to the state-of-the-art unlearning approach, we improve 5.8 comparable retrain speedup under identical settings on average on these datasets. Additionally, DeepObliviate can also pass the backdoor-based unlearning verification.

READ FULL TEXT
research
11/22/2021

Machine unlearning via GAN

Machine learning models, especially deep models, may unintentionally rem...
research
09/30/2022

Machine Unlearning Method Based On Projection Residual

Machine learning models (mainly neural networks) are used more and more ...
research
01/22/2020

A Federated Learning Framework for Privacy-preserving and Parallel Training

The deployment of such deep learning in practice has been hurdled by two...
research
02/07/2020

Machine Unlearning: Linear Filtration for Logit-based Classifiers

Recently enacted legislation grants individuals certain rights to decide...
research
07/08/2021

SSSE: Efficiently Erasing Samples from Trained Machine Learning Models

The availability of large amounts of user-provided data has been key to ...
research
06/03/2023

Forgettable Federated Linear Learning with Certified Data Removal

Federated learning (FL) is a trending distributed learning framework tha...
research
04/21/2020

Have you forgotten? A method to assess if machine learning models have forgotten data

In the era of deep learning, aggregation of data from several sources is...

Please sign up or login with your details

Forgot password? Click here to reset