No Free Lunch in "Privacy for Free: How does Dataset Condensation Help Privacy"

09/29/2022
by   Nicholas Carlini, et al.
0

New methods designed to preserve data privacy require careful scrutiny. Failure to preserve privacy is hard to detect, and yet can lead to catastrophic results when a system implementing a “privacy-preserving” method is attacked. A recent work selected for an Outstanding Paper Award at ICML 2022 (Dong et al., 2022) claims that dataset condensation (DC) significantly improves data privacy when training machine learning models. This claim is supported by theoretical analysis of a specific dataset condensation technique and an empirical evaluation of resistance to some existing membership inference attacks. In this note we examine the claims in the work of Dong et al. (2022) and describe major flaws in the empirical evaluation of the method and its theoretical analysis. These flaws imply that their work does not provide statistically significant evidence that DC improves the privacy of training ML models over a naive baseline. Moreover, previously published results show that DP-SGD, the standard approach to privacy preserving ML, simultaneously gives better accuracy and achieves a (provably) lower membership attack success rate.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/07/2021

Antipodes of Label Differential Privacy: PATE and ALIBI

We consider the privacy-preserving machine learning (ML) setting where t...
research
06/01/2022

Privacy for Free: How does Dataset Condensation Help Privacy?

To prevent unintentional data leakage, research community has resorted t...
research
08/24/2022

On Privacy Preserving Data Aggregation Protocols using BGN cryptosystem

The notion of aggregator oblivious (AO) security for privacy preserving ...
research
08/16/2021

NeuraCrypt is not private

NeuraCrypt (Yara et al. arXiv 2021) is an algorithm that converts a sens...
research
04/10/2023

Privacy-preserving Inference of Group Mean Difference in Zero-inflated Right Skewed Data with Partitioning and Censoring

We examine privacy-preserving inferences of group mean differences in ze...
research
10/07/2021

The Connection between Out-of-Distribution Generalization and Privacy of ML Models

With the goal of generalizing to out-of-distribution (OOD) data, recent ...
research
05/31/2023

A Note On Interpreting Canary Exposure

Canary exposure, introduced in Carlini et al. is frequently used to empi...

Please sign up or login with your details

Forgot password? Click here to reset