Locally Differentially Private Distributed Deep Learning via Knowledge Distillation

02/07/2022
by   Di Zhuang, et al.
0

Deep learning often requires a large amount of data. In real-world applications, e.g., healthcare applications, the data collected by a single organization (e.g., hospital) is often limited, and the majority of massive and diverse data is often segregated across multiple organizations. As such, it motivates the researchers to conduct distributed deep learning, where the data user would like to build DL models using the data segregated across multiple different data owners. However, this could lead to severe privacy concerns due to the sensitive nature of the data, thus the data owners would be hesitant and reluctant to participate. We propose LDP-DL, a privacy-preserving distributed deep learning framework via local differential privacy and knowledge distillation, where each data owner learns a teacher model using its own (local) private dataset, and the data user learns a student model to mimic the output of the ensemble of the teacher models. In the experimental evaluation, a comprehensive comparison has been made among our proposed approach (i.e., LDP-DL), DP-SGD, PATE and DP-FL, using three popular deep learning benchmark datasets (i.e., CIFAR10, MNIST and FashionMNIST). The experimental results show that LDP-DL consistently outperforms the other competitors in terms of privacy budget and model accuracy.

READ FULL TEXT
research
08/08/2019

Local Differential Privacy for Deep Learning

Deep learning (DL) is a promising area of machine learning which is beco...
research
06/05/2019

Private Deep Learning with Teacher Ensembles

Privacy-preserving deep learning is crucial for deploying deep neural ne...
research
10/12/2022

An Ensemble Teacher-Student Learning Approach with Poisson Sub-sampling to Differential Privacy Preserving Speech Recognition

We propose an ensemble learning framework with Poisson sub-sampling to e...
research
07/27/2022

Fine-grained Private Knowledge Distillation

Knowledge distillation has emerged as a scalable and effective way for p...
research
07/05/2023

SoK: Privacy-Preserving Data Synthesis

As the prevalence of data analysis grows, safeguarding data privacy has ...
research
10/26/2021

SEDML: Securely and Efficiently Harnessing Distributed Knowledge in Machine Learning

Training high-performing deep learning models require a rich amount of d...
research
08/26/2022

Federated and Privacy-Preserving Learning of Accounting Data in Financial Statement Audits

The ongoing 'digital transformation' fundamentally changes audit evidenc...

Please sign up or login with your details

Forgot password? Click here to reset