User-Entity Differential Privacy in Learning Natural Language Models

11/01/2022
by   Phung Lai, et al.
0

In this paper, we introduce a novel concept of user-entity differential privacy (UeDP) to provide formal privacy protection simultaneously to both sensitive entities in textual data and data owners in learning natural language models (NLMs). To preserve UeDP, we developed a novel algorithm, called UeDP-Alg, optimizing the trade-off between privacy loss and model utility with a tight sensitivity bound derived from seamlessly combining user and sensitive entity sampling processes. An extensive theoretical analysis and evaluation show that our UeDP-Alg outperforms baseline approaches in model utility under the same privacy budget consumption on several NLM tasks, using benchmark datasets.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/11/2022

What Does it Mean for a Language Model to Preserve Privacy?

Natural language reflects our private lives and identities, making its p...
research
08/07/2020

Privacy Guarantees for De-identifying Text Transformations

Machine Learning approaches to Natural Language Processing tasks benefit...
research
08/30/2021

Selective Differential Privacy for Language Modeling

With the increasing adoption of language models in applications involvin...
research
03/12/2021

Privacy Regularization: Joint Privacy-Utility Optimization in Language Models

Neural language models are known to have a high capacity for memorizatio...
research
09/05/2022

How Much User Context Do We Need? Privacy by Design in Mental Health NLP Application

Clinical NLP tasks such as mental health assessment from text, must take...
research
05/27/2021

On Privacy and Confidentiality of Communications in Organizational Graphs

Machine learned models trained on organizational communication data, suc...
research
07/08/2022

Bistochastic privacy

We introduce a new privacy model relying on bistochastic matrices, that ...

Please sign up or login with your details

Forgot password? Click here to reset