Privacy-Preserving Domain Adaptation of Semantic Parsers

Task-oriented dialogue systems often assist users with personal or confidential matters. For this reason, the developers of such a system are generally prohibited from observing actual usage. So how can they know where the system is failing and needs more training data or new functionality? In this work, we study ways in which realistic user utterances can be generated synthetically, to help increase the linguistic and functional coverage of the system, without compromising the privacy of actual users. To this end, we propose a two-stage Differentially Private (DP) generation method which first generates latent semantic parses, and then generates utterances based on the parses. Our proposed approach improves MAUVE by 3.8× and parse tree node-type overlap by 1.4× relative to current approaches for private synthetic data generation, improving both on fluency and semantic coverage. We further validate our approach on a realistic domain adaptation task of adding new functionality from private user data to a semantic parser, and show gains of 1.3× on its accuracy with the new feature.

READ FULL TEXT

page 8

page 10

research
07/05/2021

Differentially Private Sliced Wasserstein Distance

Developing machine learning methods that are privacy preserving is today...
research
06/15/2021

An Analysis of the Deployment of Models Trained on Private Tabular Synthetic Data: Unexpected Surprises

Diferentially private (DP) synthetic datasets are a powerful approach fo...
research
09/10/2019

Privacy-Preserving Bandits

Contextual bandit algorithms (CBAs) often rely on personal data to provi...
research
05/10/2023

Privacy-Preserving Recommender Systems with Synthetic Query Generation using Differentially Private Large Language Models

We propose a novel approach for developing privacy-preserving large-scal...
research
04/25/2023

Model Conversion via Differentially Private Data-Free Distillation

While massive valuable deep models trained on large-scale data have been...
research
12/15/2021

One size does not fit all: Investigating strategies for differentially-private learning across NLP tasks

Preserving privacy in training modern NLP models comes at a cost. We kno...
research
02/23/2022

Differentially Private Speaker Anonymization

Sharing real-world speech utterances is key to the training and deployme...

Please sign up or login with your details

Forgot password? Click here to reset