Differentially private cross-silo federated learning

07/10/2020
by   Mikko A. Heikkilä, et al.
9

Strict privacy is of paramount importance in distributed machine learning. Federated learning, with the main idea of communicating only what is needed for learning, has been recently introduced as a general approach for distributed learning to enhance learning and improve security. However, federated learning by itself does not guarantee any privacy for data subjects. To quantify and control how much privacy is compromised in the worst-case, we can use differential privacy. In this paper we combine additively homomorphic secure summation protocols with differential privacy in the so-called cross-silo federated learning setting. The goal is to learn complex models like neural networks while guaranteeing strict privacy for the individual data subjects. We demonstrate that our proposed solutions give prediction accuracy that is comparable to the non-distributed setting, and are fast enough to enable learning models with millions of parameters in a reasonable time. To enable learning under strict privacy guarantees that need privacy amplification by subsampling, we present a general algorithm for oblivious distributed subsampling. However, we also argue that when malicious parties are present, a simple approach using distributed Poisson subsampling gives better privacy. Finally, we show that by leveraging random projections we can further scale-up our approach to larger models while suffering only a modest performance loss.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
09/23/2022

Differentially private partitioned variational inference

Learning a privacy-preserving model from distributed sensitive data is a...
research
07/30/2021

Private Retrieval, Computing and Learning: Recent Progress and Future Challenges

Most of our lives are conducted in the cyberspace. The human notion of p...
research
12/10/2021

Sample and Threshold Differential Privacy: Histograms and applications

Federated analytics relies on the collection of accurate statistics abou...
research
10/13/2020

COVID-19 Imaging Data Privacy by Federated Learning Design: A Theoretical Framework

To address COVID-19 healthcare challenges, we need frequent sharing of h...
research
10/11/2021

The Skellam Mechanism for Differentially Private Federated Learning

We introduce the multi-dimensional Skellam mechanism, a discrete differe...
research
11/03/2021

Towards Sparse Federated Analytics: Location Heatmaps under Distributed Differential Privacy with Secure Aggregation

We design a scalable algorithm to privately generate location heatmaps o...
research
09/24/2021

A Generative Federated Learning Framework for Differential Privacy

In machine learning, differential privacy and federated learning concept...

Please sign up or login with your details

Forgot password? Click here to reset