Unintended memorisation of unique features in neural networks

05/20/2022
by   John Hartley, et al.
0

Neural networks pose a privacy risk due to their propensity to memorise and leak training data. We show that unique features occurring only once in training data are memorised by discriminative multi-layer perceptrons and convolutional neural networks trained on benchmark imaging datasets. We design our method for settings where sensitive training data is not available, for example medical imaging. Our setting knows the unique feature, but not the training data, model weights or the unique feature's label. We develop a score estimating a model's sensitivity to a unique feature by comparing the KL divergences of the model's output distributions given modified out-of-distribution images. We find that typical strategies to prevent overfitting do not prevent unique feature memorisation. And that images containing a unique feature are highly influential, regardless of the influence the images's other features. We also find a significant variation in memorisation with training seed. These results imply that neural networks pose a privacy risk to rarely occurring private information. This risk is more pronounced in healthcare applications since sensitive patient information can be memorised when it remains in training data due to an imperfect data sanitisation process.

READ FULL TEXT

page 2

page 4

page 16

page 17

research
02/16/2022

Measuring Unintended Memorisation of Unique Private Features in Neural Networks

Neural networks pose a privacy risk to training data due to their propen...
research
03/31/2021

Spectral decoupling allows training transferable neural networks in medical imaging

Deep neural networks show impressive performance in medical imaging task...
research
03/07/2023

Benign Overfitting for Two-layer ReLU Networks

Modern deep learning models with great expressive power can be trained t...
research
09/25/2019

A closer look at domain shift for deep learning in histopathology

Domain shift is a significant problem in histopathology. There can be la...
research
12/18/2015

Can Pretrained Neural Networks Detect Anatomy?

Convolutional neural networks demonstrated outstanding empirical results...
research
06/29/2020

Reducing Risk of Model Inversion Using Privacy-Guided Training

Machine learning models often pose a threat to the privacy of individual...

Please sign up or login with your details

Forgot password? Click here to reset