The Natural Auditor: How To Tell If Someone Used Your Words To Train Their Model

11/01/2018
by   Congzheng Song, et al.
6

To help enforce data-protection regulations such as GDPR and detect unauthorized uses of personal data, we propose a new model auditing technique that enables users to check if their data was used to train a machine learning model. We focus on auditing deep-learning models that generate natural-language text, including word prediction and dialog generation. These models are at the core of many popular online services. Furthermore, they are often trained on very sensitive personal data, such as users' messages, searches, chats, and comments. We design and evaluate an effective black-box auditing method that can detect, with very few queries to a model, if a particular user's texts were used to train it (among thousands of other users). In contrast to prior work on membership inference against ML models, we do not assume that the model produces numeric confidence values. We empirically demonstrate that we can successfully audit models that are well-generalized and not overfitted to the training data. We also analyze how text-generation models memorize word sequences and explain why this memorization makes them amenable to auditing.

READ FULL TEXT

page 1

page 4

page 11

research
06/10/2022

Membership Inference via Backdooring

Recently issued data privacy regulations like GDPR (General Data Protect...
research
09/09/2016

Stealing Machine Learning Models via Prediction APIs

Machine learning (ML) models may be deemed confidential due to their sen...
research
02/03/2020

Radioactive data: tracing through training

We want to detect whether a particular image dataset has been used to tr...
research
12/04/2021

SHAPr: An Efficient and Versatile Membership Privacy Risk Metric for Machine Learning

Data used to train machine learning (ML) models can be sensitive. Member...
research
06/17/2020

De-Anonymizing Text by Fingerprinting Language Generation

Components of machine learning systems are not (yet) perceived as securi...
research
09/17/2018

Déjà Vu: an empirical evaluation of the memorization properties of ConvNets

Convolutional neural networks memorize part of their training data, whic...
research
02/19/2020

Fawkes: Protecting Personal Privacy against Unauthorized Deep Learning Models

Today's proliferation of powerful facial recognition models poses a real...

Please sign up or login with your details

Forgot password? Click here to reset