Secure Distributed Training at Scale

06/21/2021
by   Eduard Gorbunov, et al.
5

Some of the hardest problems in deep learning can be solved with the combined effort of many independent parties, as is the case for volunteer computing and federated learning. These setups rely on high numbers of peers to provide computational resources or train on decentralized datasets. Unfortunately, participants in such systems are not always reliable. Any single participant can jeopardize the entire training run by sending incorrect updates, whether deliberately or by mistake. Training in presence of such peers requires specialized distributed training algorithms with Byzantine tolerance. These algorithms often sacrifice efficiency by introducing redundant communication or passing all updates through a trusted server. As a result, it can be infeasible to apply such algorithms to large-scale distributed deep learning, where models can have billions of parameters. In this work, we propose a novel protocol for secure (Byzantine-tolerant) decentralized training that emphasizes communication efficiency. We rigorously analyze this protocol: in particular, we provide theoretical bounds for its resistance against Byzantine and Sybil attacks and show that it has a marginal communication overhead. To demonstrate its practical effectiveness, we conduct large-scale experiments on image classification and language modeling in presence of Byzantine attackers.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/21/2021

Bristle: Decentralized Federated Learning in Byzantine, Non-i.i.d. Environments

Federated learning (FL) is a privacy-friendly type of machine learning w...
research
02/20/2020

Towards Byzantine-resilient Learning in Decentralized Systems

With the proliferation of IoT and edge computing, decentralized learning...
research
06/08/2020

Secure Byzantine-Robust Machine Learning

Increasingly machine learning systems are being deployed to edge servers...
research
10/04/2022

Shielding Federated Learning: Mitigating Byzantine Attacks with Less Constraints

Federated learning is a newly emerging distributed learning framework th...
research
03/07/2023

Can Decentralized Learning be more robust than Federated Learning?

Decentralized Learning (DL) is a peer–to–peer learning approach that all...
research
10/14/2019

Election Coding for Distributed Learning: Protecting SignSGD against Byzantine Attacks

Recent advances in large-scale distributed learning algorithms have enab...
research
11/12/2020

Analysis of the Matrix Event Graph Replicated Data Type

Matrix is a new kind of decentralized, topic-based publish-subscribe mid...

Please sign up or login with your details

Forgot password? Click here to reset