Archimedes Meets Privacy: On Privately Estimating Quantiles in High Dimensions Under Minimal Assumptions

08/15/2022
by   Omri Ben-Eliezer, et al.
0

The last few years have seen a surge of work on high dimensional statistics under privacy constraints, mostly following two main lines of work: the “worst case” line, which does not make any distributional assumptions on the input data; and the “strong assumptions” line, which assumes that the data is generated from specific families, e.g., subgaussian distributions. In this work we take a middle ground, obtaining new differentially private algorithms with polynomial sample complexity for estimating quantiles in high-dimensions, as well as estimating and sampling points of high Tukey depth, all working under very mild distributional assumptions. From the technical perspective, our work relies upon deep robustness results in the convex geometry literature, demonstrating how such results can be used in a private context. Our main object of interest is the (convex) floating body (FB), a notion going back to Archimedes, which is a robust and well studied high-dimensional analogue of the interquantile range. We show how one can privately, and with polynomially many samples, (a) output an approximate interior point of the FB – e.g., “a typical user” in a high-dimensional database – by leveraging the robustness of the Steiner point of the FB; and at the expense of polynomially many more samples, (b) produce an approximate uniform sample from the FB, by constructing a private noisy projection oracle.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/01/2018

Privately Learning High-Dimensional Distributions

We design nearly optimal differentially private algorithms for learning ...
research
02/27/2019

Private Center Points and Learning of Halfspaces

We present a private learner for halfspaces over an arbitrary finite dom...
research
12/09/2022

Robustness Implies Privacy in Statistical Estimation

We study the relationship between adversarial robustness and differentia...
research
11/12/2020

Optimal Private Median Estimation under Minimal Distributional Assumptions

We study the fundamental task of estimating the median of an underlying ...
research
09/09/2019

Differentially Private Algorithms for Learning Mixtures of Separated Gaussians

Learning the parameters of a Gaussian mixtures models is a fundamental a...
research
10/08/2020

Online and Distribution-Free Robustness: Regression and Contextual Bandits with Huber Contamination

In this work we revisit two classic high-dimensional online learning pro...
research
02/16/2013

k-d Darts: Sampling by k-Dimensional Flat Searches

We formalize the notion of sampling a function using k-d darts. A k-d da...

Please sign up or login with your details

Forgot password? Click here to reset