Towards Robust and Privacy-preserving Text Representations

05/16/2018
by   Yitong Li, et al.
0

Written text often provides sufficient clues to identify the author, their gender, age, and other important attributes. Consequently, the authorship of training and evaluation corpora can have unforeseen impacts, including differing model performance for different user groups, as well as privacy implications. In this paper, we propose an approach to explicitly obscure important author characteristics at training time, such that representations learned are invariant to these attributes. Evaluating on two tasks, we show that this leads to increased privacy in the learned representations, as well as more robust models to varying evaluation conditions, including out-of-domain corpora.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
11/06/2017

A^4NT: Author Attribute Anonymity by Adversarial Training of Neural Machine Translation

Text-based analysis methods allow to reveal privacy relevant author attr...
research
05/12/2022

Fair NLP Models with Differentially Private Text Encoders

Encoded text representations often capture sensitive attributes about in...
research
05/28/2019

Overlearning Reveals Sensitive Attributes

`Overlearning' means that a model trained for a seemingly simple objecti...
research
09/02/2020

Too good to be true? Predicting author profiles from abusive language

The problem of online threats and abuse could potentially be mitigated w...
research
06/19/2018

Private Text Classification

Confidential text corpora exist in many forms, but do not allow arbitrar...
research
07/12/2017

The Case for Being Average: A Mediocrity Approach to Style Masking and Author Obfuscation

Users posting online expect to remain anonymous unless they have logged ...
research
06/24/2021

Learning Language and Multimodal Privacy-Preserving Markers of Mood from Mobile Data

Mental health conditions remain underdiagnosed even in countries with co...

Please sign up or login with your details

Forgot password? Click here to reset