The Secret Sharer: Measuring Unintended Neural Network Memorization & Extracting Secrets

02/22/2018
by   Nicholas Carlini, et al.
0

Machine learning models based on neural networks and deep learning are being rapidly adopted for many purposes. What those models learn, and what they may share, is a significant concern when the training data may contain secrets and the models are public -- e.g., when a model helps users compose text messages using models trained on all users' messages. This paper presents exposure: a simple-to-compute metric that can be applied to any deep learning model for measuring the memorization of secrets. Using this metric, we show how to extract those secrets efficiently using black-box API access. Further, we show that unintended memorization occurs early, is not due to over-fitting, and is a persistent issue across different types of models, hyperparameters, and training strategies. We experiment with both real-world models (e.g., a state-of-the-art translation model) and datasets (e.g., the Enron email dataset, which contains users' credit card numbers) to demonstrate both the utility of measuring exposure and the ability to extract secrets. Finally, we consider many defenses, finding some ineffective (like regularization), and others to lack guarantees. However, by instantiating our own differentially-private recurrent model, we validate that by appropriately investing in the use of state-of-the-art techniques, the problem can be resolved, with high utility.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
12/31/2019

Hiding Information in Big Data based on Deep Learning

The current approach of information hiding based on deep learning model ...
research
05/22/2019

A framework for the extraction of Deep Neural Networks by leveraging public data

Machine learning models trained on confidential datasets are increasingl...
research
03/02/2021

DPlis: Boosting Utility of Differentially Private Deep Learning via Randomized Smoothing

Deep learning techniques have achieved remarkable performance in wide-ra...
research
07/12/2021

Improving the Algorithm of Deep Learning with Differential Privacy

In this paper, an adjustment to the original differentially private stoc...
research
06/24/2022

"You Can't Fix What You Can't Measure": Privately Measuring Demographic Performance Disparities in Federated Learning

Federated learning allows many devices to collaborate in the training of...

Please sign up or login with your details

Forgot password? Click here to reset