Privacy-preserving Distributed Machine Learning via Local Randomization and ADMM Perturbation

07/30/2019
by   Xin Wang, et al.
0

With the proliferation of training data, distributed machine learning (DML) is becoming more competent for large-scale learning tasks. However, privacy concern has to be attached prior importance in DML, since training data may contain sensitive information of users. Most existing privacy-aware schemes are established based on an assumption that the users trust the server collecting their data, and are limited to provide the same privacy guarantee for the entire data sample. In this paper, we remove the trustworthy servers assumption, and propose a privacy-preserving ADMM-based DML framework that preserves heterogeneous privacy for users' data. The new challenging issue is to reduce the accumulation of privacy losses over ADMM iterations as much as possible. In the proposed privacy-aware DML framework, a local randomization approach, which is proved to be differentially private, is adopted to provide users with self-controlled privacy guarantee for the most sensitive information. Further, the ADMM algorithm is perturbed through a combined noise-adding method, which simultaneously preserves privacy for users' less sensitive information and strengthens the privacy protection of the most sensitive information. Also, we analyze the performance of the trained model according to its generalization error. Finally, we conduct extensive experiments using synthetic and real-world datasets to validate the theoretical results and evaluate the classification performance of the proposed framework.

READ FULL TEXT

page 1

page 13

research
01/07/2019

Optimal Differentially Private ADMM for Distributed Machine Learning

Due to massive amounts of data distributed across multiple locations, di...
research
02/16/2019

On Privacy-preserving Decentralized Optimization through Alternating Direction Method of Multipliers

Privacy concerns with sensitive data in machine learning are receiving i...
research
02/10/2018

Distributed One-class Learning

We propose a cloud-based filter trained to block third parties from uplo...
research
08/11/2020

Towards Plausible Differentially Private ADMM Based Distributed Machine Learning

The Alternating Direction Method of Multipliers (ADMM) and its distribut...
research
02/06/2020

Privacy-Preserving Boosting in the Local Setting

In machine learning, boosting is one of the most popular methods that de...
research
07/24/2017

Share your Model instead of your Data: Privacy Preserving Mimic Learning for Ranking

Deep neural networks have become a primary tool for solving problems in ...
research
01/08/2019

Data Masking with Privacy Guarantees

We study the problem of data release with privacy, where data is made av...

Please sign up or login with your details

Forgot password? Click here to reset