Private Distribution Testing with Heterogeneous Constraints: Your Epsilon Might Not Be Mine

09/12/2023
by   Clément L. Canonne, et al.
0

Private closeness testing asks to decide whether the underlying probability distributions of two sensitive datasets are identical or differ significantly in statistical distance, while guaranteeing (differential) privacy of the data. As in most (if not all) distribution testing questions studied under privacy constraints, however, previous work assumes that the two datasets are equally sensitive, i.e., must be provided the same privacy guarantees. This is often an unrealistic assumption, as different sources of data come with different privacy requirements; as a result, known closeness testing algorithms might be unnecessarily conservative, "paying" too high a privacy budget for half of the data. In this work, we initiate the study of the closeness testing problem under heterogeneous privacy constraints, where the two datasets come with distinct privacy requirements. We formalize the question and provide algorithms under the three most widely used differential privacy settings, with a particular focus on the local and shuffle models of privacy; and show that one can indeed achieve better sample efficiency when taking into account the two different "epsilon" requirements.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
08/20/2021

Uniformity Testing in the Shuffle Model: Simpler, Better, Faster

Uniformity testing, or testing whether independent observations are unif...
research
02/21/2020

Locally Private Hypothesis Selection

We initiate the study of hypothesis selection under local differential p...
research
08/11/2021

Statistical Inference in the Differential Privacy Model

In modern settings of data analysis, we may be running our algorithms on...
research
11/07/2022

Lessons Learned: Surveying the Practicality of Differential Privacy in the Industry

Since its introduction in 2006, differential privacy has emerged as a pr...
research
12/15/2018

A General Approach to Adding Differential Privacy to Iterative Training Procedures

In this work we address the practical challenges of training machine lea...
research
06/07/2022

Histogram Estimation under User-level Privacy with Heterogeneous Data

We study the problem of histogram estimation under user-level differenti...
research
02/23/2015

Learning with Differential Privacy: Stability, Learnability and the Sufficiency and Necessity of ERM Principle

While machine learning has proven to be a powerful data-driven solution ...

Please sign up or login with your details

Forgot password? Click here to reset