Differentially Private Histograms in the Shuffle Model from Fake Users

04/06/2021
by   Albert Cheu, et al.
0

There has been much recent work in the shuffle model of differential privacy, particularly for approximate d-bin histograms. While these protocols achieve low error, the number of messages sent by each user – the message complexity – has so far scaled with d or the privacy parameters. The message complexity is an informative predictor of a shuffle protocol's resource consumption. We present a protocol whose message complexity is two when there are sufficiently many users. The protocol essentially pairs each row in the dataset with a fake row and performs a simple randomization on all rows. We show that the error introduced by the protocol is small, using rigorous analysis as well as experiments on real-world data. We also prove that corrupt users have a relatively low impact on our protocol's estimates.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
11/12/2021

Frequency Estimation in the Shuffle Model with (Almost) a Single Message

We present a protocol in the shuffle model of differential privacy for t...
research
12/19/2021

Pure Differential Privacy from Secure Intermediaries

Recent work in differential privacy has explored the prospect of combini...
research
02/05/2020

Pure Differentially Private Summation from Anonymous Messages

The shuffled (aka anonymous) model has recently generated significant in...
research
09/21/2020

On Distributed Differential Privacy and Counting Distinct Elements

We study the setup where each of n users holds an element from a discret...
research
09/27/2021

Differentially Private Aggregation in the Shuffle Model: Almost Central Accuracy in Almost a Single Message

The shuffle model of differential privacy has attracted attention in the...
research
10/29/2020

Differential Privacy and Natural Language Processing to Generate Contextually Similar Decoy Messages in Honey Encryption Scheme

Honey Encryption is an approach to encrypt the messages using low min-en...
research
06/25/2020

Towards Differentially Private Text Representations

Most deep learning frameworks require users to pool their local data or ...

Please sign up or login with your details

Forgot password? Click here to reset