ADePT: Auto-encoder based Differentially Private Text Transformation

01/29/2021
by   Satyapriya Krishna, et al.
7

Privacy is an important concern when building statistical models on data containing personal information. Differential privacy offers a strong definition of privacy and can be used to solve several privacy concerns (Dwork et al., 2014). Multiple solutions have been proposed for the differentially-private transformation of datasets containing sensitive information. However, such transformation algorithms offer poor utility in Natural Language Processing (NLP) tasks due to noise added in the process. In this paper, we address this issue by providing a utility-preserving differentially private text transformation algorithm using auto-encoders. Our algorithm transforms text to offer robustness against attacks and produces transformations with high semantic quality that perform well on downstream NLP tasks. We prove the theoretical privacy guarantee of our algorithm and assess its privacy leakage under Membership Inference Attacks(MIA) (Shokri et al., 2017) on models trained with transformed data. Our results show that the proposed model performs better against MIA attacks while offering lower to no degradation in the utility of the underlying transformation process compared to existing baselines.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
09/07/2021

When differential privacy meets NLP: The devil is in the detail

Differential privacy provides a formal approach to privacy of individual...
research
07/20/2019

ER-AE: Differentially-private Text Generation for Authorship Anonymization

Most of privacy protection studies for textual data focus on removing ex...
research
05/31/2023

Adaptive False Discovery Rate Control with Privacy Guarantee

Differentially private multiple testing procedures can protect the infor...
research
03/04/2021

On the privacy-utility trade-off in differentially private hierarchical text classification

Hierarchical models for text classification can leak sensitive or confid...
research
02/24/2022

How reparametrization trick broke differentially-private text representation learning

As privacy gains traction in the NLP community, researchers have started...
research
05/12/2023

Differentially Private Set-Based Estimation Using Zonotopes

For large-scale cyber-physical systems, the collaboration of spatially d...
research
06/01/2022

Privacy for Free: How does Dataset Condensation Help Privacy?

To prevent unintentional data leakage, research community has resorted t...

Please sign up or login with your details

Forgot password? Click here to reset