Analyzing Privacy Loss in Updates of Natural Language Models

12/17/2019
by   Shruti Tople, et al.
0

To continuously improve quality and reflect changes in data, machine learning-based services have to regularly re-train and update their core models. In the setting of language models, we show that a comparative analysis of model snapshots before and after an update can reveal a surprising amount of detailed information about the changes in the data used for training before and after the update. We discuss the privacy implications of our findings, propose mitigation strategies and evaluate their effect.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/11/2022

What Does it Mean for a Language Model to Preserve Privacy?

Natural language reflects our private lives and identities, making its p...
research
04/25/2020

Learning to Update Natural Language Comments Based on Code Changes

We formulate the novel task of automatically updating an existing natura...
research
10/23/2020

Overcoming Conflicting Data for Model Updates

In this paper, we explore how to use a small amount of new data to updat...
research
03/12/2021

Privacy Regularization: Joint Privacy-Utility Optimization in Language Models

Neural language models are known to have a high capacity for memorizatio...
research
10/13/2022

Is It Worth the (Environmental) Cost? Limited Evidence for the Benefits of Diachronic Continuous Training

Language is constantly changing and evolving, leaving language models to...
research
07/12/2023

A Comprehensive Overview of Large Language Models

Large Language Models (LLMs) have shown excellent generalization capabil...
research
04/19/2023

A Theory on Adam Instability in Large-Scale Machine Learning

We present a theory for the previously unexplained divergent behavior no...

Please sign up or login with your details

Forgot password? Click here to reset