Ultimate Power of Inference Attacks: Privacy Risks of High-Dimensional Models

05/29/2019
by   Sasi Kumar Murakonda, et al.
0

Models leak information about their training data. This enables attackers to infer sensitive information about their training sets, notably determine if a data sample was part of the model's training set. The existing works empirically show the possibility of these tracing (membership inference) attacks against complex models with a large number of parameters. However, the attack results are dependent on the specific training data, can be obtained only after the tedious process of training the model and performing the attack, and are missing any measure of the confidence and unused potential power of the attack. A model designer is interested in identifying which model structures leak more information, how adding new parameters to the model increases its privacy risk, and what is the gain of adding new data points to decrease the overall information leakage. The privacy analysis should also enable designing the most powerful inference attack. In this paper, we design a theoretical framework to analyze the maximum power of tracing attacks against high-dimensional models, with the focus on probabilistic graphical models. We provide a tight upper-bound on the power (true positive rate) of these attacks, with respect to their error (false positive rate). The bound, as it should be, is independent of the knowledge and algorithm of any specific attack, as well as the values of particular samples in the training set. It provides a measure of the potential leakage of a model given its structure, as a function of the structure complexity and the size of training set.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
07/16/2018

Machine Learning with Membership Privacy using Adversarial Regularization

Machine learning models leak information about the datasets on which the...
research
12/06/2022

On the Discredibility of Membership Inference Attacks

With the wide-spread application of machine learning models, it has beco...
research
06/09/2020

On the Effectiveness of Regularization Against Membership Inference Attacks

Deep learning models often raise privacy concerns as they leak informati...
research
02/21/2019

Learning requirements for stealth attacks

The learning data requirements are analyzed for the construction of stea...
research
09/18/2022

Distribution inference risks: Identifying and mitigating sources of leakage

A large body of work shows that machine learning (ML) models can leak se...
research
01/28/2021

An Analysis Of Protected Health Information Leakage In Deep-Learning Based De-Identification Algorithms

The increasing complexity of algorithms for analyzing medical data, incl...
research
05/09/2021

Bounding Information Leakage in Machine Learning

Machine Learning services are being deployed in a large range of applica...

Please sign up or login with your details

Forgot password? Click here to reset