Just Fine-tune Twice: Selective Differential Privacy for Large Language Models

04/15/2022
by   Weiyan Shi, et al.
0

With the increasing adoption of NLP models in real-world products, it becomes more and more important to protect these models from privacy leakage. Because private information in language data is sparse, previous research formalized a Selective-Differential-Privacy (SDP) notion to provide protection for sensitive tokens detected by policy functions, and prove its effectiveness on RNN-based models. But the previous mechanism requires separating the private and public model parameters and thus cannot be applied on large attention-based models. In this paper, we propose a simple yet effective just-fine-tune-twice privacy mechanism to first fine-tune on in-domain redacted data and then on in-domain private data, to achieve SDP for large Transformer-based language models. We also design explicit and contextual policy functions to provide protections at different levels. Experiments show that our models achieve strong performance while staying robust to the canary insertion attack. We further show that even under low-resource settings with a small amount of in-domain data, SDP can still improve the model utility. We will release the code, data and models to facilitate future research.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
08/30/2021

Selective Differential Privacy for Language Modeling

With the increasing adoption of language models in applications involvin...
research
05/02/2023

Mitigating Approximate Memorization in Language Models via Dissimilarity Learned Policy

Large Language models (LLMs) are trained on large amounts of data, which...
research
05/28/2023

Training Private Models That Know What They Don't Know

Training reliable deep learning models which avoid making overconfident ...
research
01/29/2021

N-grams Bayesian Differential Privacy

Differential privacy has gained popularity in machine learning as a stro...
research
07/04/2022

A Customised Text Privatisation Mechanism with Differential Privacy

In Natural Language Understanding (NLU) applications, training an effect...
research
04/07/2023

Does Prompt-Tuning Language Model Ensure Privacy?

Prompt-tuning has received attention as an efficient tuning method in th...
research
09/06/2023

Hide and Seek (HaS): A Lightweight Framework for Prompt Privacy Protection

Numerous companies have started offering services based on large languag...

Please sign up or login with your details

Forgot password? Click here to reset