Measuring Data Leakage in Machine-Learning Models with Fisher Information

02/23/2021
by   Awni Hannun, et al.
0

Machine-learning models contain information about the data they were trained on. This information leaks either through the model itself or through predictions made by the model. Consequently, when the training data contains sensitive attributes, assessing the amount of information leakage is paramount. We propose a method to quantify this leakage using the Fisher information of the model about the data. Unlike the worst-case a priori guarantees of differential privacy, Fisher information loss measures leakage with respect to specific examples, attributes, or sub-populations within the dataset. We motivate Fisher information loss through the Cramér-Rao bound and delineate the implied threat model. We provide efficient methods to compute Fisher information loss for output-perturbed generalized linear models. Finally, we empirically validate Fisher information loss as a useful measure of information leakage.

READ FULL TEXT
research
09/21/2022

Measuring and Controlling Split Layer Privacy Leakage Using Fisher Information

Split learning and inference propose to run training/inference of a larg...
research
01/28/2022

Bounding Training Data Reconstruction in Private (Deep) Learning

Differential privacy is widely accepted as the de facto method for preve...
research
10/24/2022

Explaining epsilon in differential privacy through the lens of information theory

The study of leakage measures for privacy has been a subject of intensiv...
research
05/09/2021

Bounding Information Leakage in Machine Learning

Machine Learning services are being deployed in a large range of applica...
research
03/26/2018

On the loss of Fisher information in some multi-object tracking observation models

The concept of Fisher information can be useful even in cases where the ...
research
12/28/2020

Implementation of Security Systems for Detection and Prevention of Data Loss/Leakage at Organization via Traffic Inspection

Data Loss/Leakage Prevention (DLP) continues to be the main issue for ma...
research
03/17/2022

Leveraging Adversarial Examples to Quantify Membership Information Leakage

The use of personal data for training machine learning systems comes wit...

Please sign up or login with your details

Forgot password? Click here to reset