Leveraging Algorithmic Fairness to Mitigate Blackbox Attribute Inference Attacks

11/18/2022
by   Jan Aalmoes, et al.
0

Machine learning (ML) models have been deployed for high-stakes applications, e.g., healthcare and criminal justice. Prior work has shown that ML models are vulnerable to attribute inference attacks where an adversary, with some background knowledge, trains an ML attack model to infer sensitive attributes by exploiting distinguishable model predictions. However, some prior attribute inference attacks have strong assumptions about adversary's background knowledge (e.g., marginal distribution of sensitive attribute) and pose no more privacy risk than statistical inference. Moreover, none of the prior attacks account for class imbalance of sensitive attribute in datasets coming from real-world applications (e.g., Race and Sex). In this paper, we propose an practical and effective attribute inference attack that accounts for this imbalance using an adaptive threshold over the attack model's predictions. We exhaustively evaluate our proposed attack on multiple datasets and show that the adaptive threshold over the model's predictions drastically improves the attack accuracy over prior work. Finally, current literature lacks an effective defence against attribute inference attacks. We investigate the impact of fairness constraints (i.e., designed to mitigate unfairness in model predictions) during model training on our attribute inference attack. We show that constraint based fairness algorithms which enforces equalized odds acts as an effective defense against attribute inference attacks without impacting the model utility. Hence, the objective of algorithmic fairness and sensitive attribute privacy are aligned.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/04/2022

Dikaios: Privacy Auditing of Algorithmic Fairness via Attribute Inference Attacks

Machine learning (ML) models have been deployed for high-stakes applicat...
research
08/21/2022

Inferring Sensitive Attributes from Model Explanations

Model explanations provide transparency into a trained machine learning ...
research
02/27/2022

Attacks on Deidentification's Defenses

Quasi-identifier-based deidentification techniques (QI-deidentification)...
research
09/29/2018

Statistical Inference Attack Against PHY-layer Key Extraction and Countermeasures

The formal theoretical analysis on channel correlations in both real ind...
research
12/12/2019

Awareness in Practice: Tensions in Access to Sensitive Attribute Data for Antidiscrimination

Organizations cannot address demographic disparities that they cannot se...
research
06/19/2019

Adversarial Task-Specific Privacy Preservation under Attribute Attack

With the prevalence of machine learning services, crowdsourced data cont...
research
10/31/2020

Evaluation of Inference Attack Models for Deep Learning on Medical Data

Deep learning has attracted broad interest in healthcare and medical com...

Please sign up or login with your details

Forgot password? Click here to reset