Unleashing the Power of Randomization in Auditing Differentially Private ML

05/29/2023
by   Krishna Pillutla, et al.
0

We present a rigorous methodology for auditing differentially private machine learning algorithms by adding multiple carefully designed examples called canaries. We take a first principles approach based on three key components. First, we introduce Lifted Differential Privacy (LiDP) that expands the definition of differential privacy to handle randomized datasets. This gives us the freedom to design randomized canaries. Second, we audit LiDP by trying to distinguish between the model trained with K canaries versus K - 1 canaries in the dataset, leaving one canary out. By drawing the canaries i.i.d., LiDP can leverage the symmetry in the design and reuse each privately trained model to run multiple statistical tests, one for each canary. Third, we introduce novel confidence intervals that take advantage of the multiple test statistics by adapting to the empirical higher-order correlations. Together, this new recipe demonstrates significant improvements in sample complexity, both theoretically and empirically, using synthetic and real data. Further, recent advances in designing stronger canaries can be readily incorporated into the new framework.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
04/11/2018

Differentially Private Confidence Intervals for Empirical Risk Minimization

The process of data mining with differential privacy produces results th...
research
09/12/2023

Deciding Differential Privacy of Online Algorithms with Multiple Variables

We consider the problem of checking the differential privacy of online r...
research
09/05/2019

Duet: An Expressive Higher-order Language and Linear Type System for Statically Enforcing Differential Privacy

During the past decade, differential privacy has become the gold standar...
research
06/10/2022

Bayesian Estimation of Differential Privacy

Algorithms such as Differentially Private SGD enable training machine le...
research
08/01/2018

A Differentially Private Kernel Two-Sample Test

Kernel two-sample testing is a useful statistical tool in determining wh...
research
08/10/2018

Ektelo: A Framework for Defining Differentially-Private Computations

The adoption of differential privacy is growing but the complexity of de...
research
08/21/2021

Statistical Quantification of Differential Privacy: A Local Approach

In this work we introduce a new approach for statistical quantification ...

Please sign up or login with your details

Forgot password? Click here to reset