An Empirical Study on the Membership Inference Attack against Tabular Data Synthesis Models

08/17/2022
by   Jihyeon Hyeong, et al.
0

Tabular data typically contains private and important information; thus, precautions must be taken before they are shared with others. Although several methods (e.g., differential privacy and k-anonymity) have been proposed to prevent information leakage, in recent years, tabular data synthesis models have become popular because they can well trade-off between data utility and privacy. However, recent research has shown that generative models for image data are susceptible to the membership inference attack, which can determine whether a given record was used to train a victim synthesis model. In this paper, we investigate the membership inference attack in the context of tabular data synthesis. We conduct experiments on 4 state-of-the-art tabular data synthesis models under two attack scenarios (i.e., one black-box and one white-box attack), and find that the membership inference attack can seriously jeopardize these models. We next conduct experiments to evaluate how well two popular differentially-private deep learning training algorithms, DP-SGD and DP-GAN, can protect the models against the attack. Our key finding is that both algorithms can largely alleviate this threat by sacrificing the generation quality.

READ FULL TEXT
research
12/06/2018

Differentially Private Data Generative Models

Deep neural networks (DNNs) have recently been widely adopted in various...
research
02/07/2023

Membership Inference Attacks against Diffusion Models

Diffusion models have attracted attention in recent years as innovative ...
research
04/16/2021

Membership Inference Attack Susceptibility of Clinical Language Models

Deep Neural Network (DNN) models have been shown to have high empirical ...
research
09/06/2023

Hide and Seek (HaS): A Lightweight Framework for Prompt Privacy Protection

Numerous companies have started offering services based on large languag...
research
03/04/2021

On the privacy-utility trade-off in differentially private hierarchical text classification

Hierarchical models for text classification can leak sensitive or confid...
research
02/24/2017

Deep Models Under the GAN: Information Leakage from Collaborative Deep Learning

Deep Learning has recently become hugely popular in machine learning, pr...
research
08/23/2023

A Probabilistic Fluctuation based Membership Inference Attack for Diffusion Models

Membership Inference Attack (MIA) identifies whether a record exists in ...

Please sign up or login with your details

Forgot password? Click here to reset