Every Query Counts: Analyzing the Privacy Loss of Exploratory Data Analyses

08/27/2020
by   Saskia Nuñez von Voigt, et al.
0

An exploratory data analysis is an essential step for every data analyst to gain insights, evaluate data quality and (if required) select a machine learning model for further processing. While privacy-preserving machine learning is on the rise, more often than not this initial analysis is not counted towards the privacy budget. In this paper, we quantify the privacy loss for basic statistical functions and highlight the importance of taking it into account when calculating the privacy-loss budget of a machine learning approach.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/24/2019

When Relaxations Go Bad: "Differentially-Private" Machine Learning

Differential privacy is becoming a standard notion for performing privac...
research
03/27/2018

Privacy Preserving Machine Learning: Threats and Solutions

For privacy concerns to be addressed adequately in current machine learn...
research
11/11/2019

Privacy-Preserving Gradient Boosting Decision Trees

The Gradient Boosting Decision Tree (GBDT) is a popular machine learning...
research
07/10/2023

Privacy-Preserving Graph Machine Learning from Data to Computation: A Survey

In graph machine learning, data collection, sharing, and analysis often ...
research
03/18/2020

Predicting Performance of Asynchronous Differentially-Private Learning

We consider training machine learning models using Training data located...
research
11/16/2015

How much does your data exploration overfit? Controlling bias via information usage

Modern data is messy and high-dimensional, and it is often not clear a p...
research
03/30/2022

Geographic Spines in the 2020 Census Disclosure Avoidance System TopDown Algorithm

The TopDown Algorithm (TDA) first produces differentially private counts...

Please sign up or login with your details

Forgot password? Click here to reset