When differential privacy meets NLP: The devil is in the detail

09/07/2021
by   Ivan Habernal, et al.
0

Differential privacy provides a formal approach to privacy of individuals. Applications of differential privacy in various scenarios, such as protecting users' original utterances, must satisfy certain mathematical properties. Our contribution is a formal analysis of ADePT, a differentially private auto-encoder for text rewriting (Krishna et al, 2021). ADePT achieves promising results on downstream tasks while providing tight privacy guarantees. Our proof reveals that ADePT is not differentially private, thus rendering the experimental results unsubstantiated. We also quantify the impact of the error in its private mechanism, showing that the true sensitivity is higher by at least factor 6 in an optimistic case of a very small encoder's dimension and that the amount of utterances that are not privatized could easily reach 100 of the entire dataset. Our intention is neither to criticize the authors, nor the peer-reviewing process, but rather point out that if differential privacy applications in NLP rely on formal guarantees, these should be outlined in full and put under detailed scrutiny.

READ FULL TEXT
research
02/24/2022

How reparametrization trick broke differentially-private text representation learning

As privacy gains traction in the NLP community, researchers have started...
research
01/29/2021

ADePT: Auto-encoder based Differentially Private Text Transformation

Privacy is an important concern when building statistical models on data...
research
05/12/2022

Fair NLP Models with Differentially Private Text Encoders

Encoded text representations often capture sensitive attributes about in...
research
02/10/2020

Guidelines for Implementing and Auditing Differentially Private Systems

Differential privacy is an information theoretic constraint on algorithm...
research
07/04/2018

Privacy Amplification by Subsampling: Tight Analyses via Couplings and Divergences

Differential privacy comes equipped with multiple analytical tools for t...
research
05/29/2019

Fuzzi: A Three-Level Logic for Differential Privacy

Curators of sensitive datasets sometimes need to know whether queries ag...
research
10/28/2019

Differentially Private Distributed Data Summarization under Covariate Shift

We envision AI marketplaces to be platforms where consumers, with very l...

Please sign up or login with your details

Forgot password? Click here to reset