Protecting Global Properties of Datasets with Distribution Privacy Mechanisms

07/18/2022
by   Michelle Chen, et al.
0

Alongside the rapid development of data collection and analysis techniques in recent years, there is increasingly an emphasis on the need to address information leakage associated with such usage of data. To this end, much work in the privacy literature is devoted to the protection of individual users and contributors of data. However, some situations instead require a different notion of data confidentiality involving global properties aggregated over the records of a dataset. Such notions of information protection are particularly applicable for business and organization data, where global properties may reflect trade secrets, or demographic data, which can be harmful if mishandled. Recent work on property inference attacks furthermore shows how data analysis algorithms can be susceptible to leaking these global properties of data, highlighting the importance of developing mechanisms that can protect such information. In this work, we demonstrate how a distribution privacy framework can be applied to formalize the problem of protecting global properties of datasets. Given this framework, we investigate several mechanisms and their tradeoffs for providing this notion of data confidentiality. We analyze the theoretical protection guarantees offered by these mechanisms under various data assumptions, then implement and empirically evaluate these mechanisms for several data analysis tasks. The results of our experiments show that our mechanisms can indeed reduce the effectiveness of practical property inference attacks while providing utility substantially greater than a crude group differential privacy baseline. Our work thus provides groundwork for theoretically supported mechanisms for protecting global properties of datasets.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/22/2023

Analyzing the Shuffle Model through the Lens of Quantitative Information Flow

Local differential privacy (LDP) is a variant of differential privacy (D...
research
11/04/2019

Providing Input-Discriminative Protection for Local Differential Privacy

Local Differential Privacy (LDP) provides provable privacy protection fo...
research
07/02/2017

Privacy-Preserving Mechanisms for Parametric Survival Analysis with Weibull Distribution

Survival analysis studies the statistical properties of the time until a...
research
02/01/2021

Regionalized location obfuscation mechanism with personalized privacy levels

Global Positioning Systems are now a standard module in mobile devices, ...
research
10/30/2019

Chasing Accuracy and Privacy, and Catching Both: A Literature Survey on Differentially Private Histogram Publication

Histograms and synthetic data are of key importance in data analysis. Ho...
research
07/15/2019

Single-Component Privacy Guarantees in Helper Data Systems and Sparse Coding

We investigate the privacy of two approaches to (biometric) template pro...
research
03/22/2021

Privacy-aware Process Performance Indicators: Framework and Release Mechanisms

Process performance indicators (PPIs) are metrics to quantify the degree...

Please sign up or login with your details

Forgot password? Click here to reset