Selective Pre-training for Private Fine-tuning

05/23/2023
by   Da Yu, et al.
0

Suppose we want to train text prediction models in email clients or word processors. The models must preserve the privacy of user data and adhere to a specific fixed size to meet memory and inference time requirements. We introduce a generic framework to solve this problem. Specifically, we are given a public dataset D_pub and a private dataset D_priv corresponding to a downstream task T. How should we pre-train a fixed-size model M on D_pub and fine-tune it on D_priv such that performance of M with respect to T is maximized and M satisfies differential privacy with respect to D_priv? We show that pre-training on a subset of dataset D_pub that brings the public distribution closer to the private distribution is a crucial ingredient to maximize the transfer learning abilities of M after pre-training, especially in the regimes where model sizes are relatively small. Besides performance improvements, our framework also shows that with careful pre-training and private fine-tuning, smaller models can match the performance of much larger models, highlighting the promise of differentially private training as a tool for model compression and efficiency.

READ FULL TEXT
research
09/12/2023

Exploring the Benefits of Differentially Private Pre-training and Parameter-Efficient Fine-tuning for Table Transformers

For machine learning with tabular data, Table Transformer (TabTransforme...
research
07/25/2023

Accuracy Amplification in Differentially Private Logistic Regression: A Pre-Training Approach

Machine learning (ML) models can memorize training datasets. As a result...
research
10/02/2020

Data Transfer Approaches to Improve Seq-to-Seq Retrosynthesis

Retrosynthesis is a problem to infer reactant compounds to synthesize a ...
research
03/22/2022

Mixed Differential Privacy in Computer Vision

We introduce AdaMix, an adaptive differentially private algorithm for tr...
research
02/27/2023

Differentially Private Diffusion Models Generate Useful Synthetic Images

The ability to generate privacy-preserving synthetic versions of sensiti...
research
05/24/2023

Flocks of Stochastic Parrots: Differentially Private Prompt Learning for Large Language Models

Large language models (LLMs) are excellent in-context learners. However,...
research
05/26/2022

Differentially Private Decoding in Large Language Models

Recent large-scale natural language processing (NLP) systems use a pre-t...

Please sign up or login with your details

Forgot password? Click here to reset