Papaya: Practical, Private, and Scalable Federated Learning

11/08/2021
by   Dzmitry Huba, et al.
0

Cross-device Federated Learning (FL) is a distributed learning paradigm with several challenges that differentiate it from traditional distributed learning, variability in the system characteristics on each device, and millions of clients coordinating with a central server being primary ones. Most FL systems described in the literature are synchronous - they perform a synchronized aggregation of model updates from individual clients. Scaling synchronous FL is challenging since increasing the number of clients training in parallel leads to diminishing returns in training speed, analogous to large-batch training. Moreover, stragglers hinder synchronous FL training. In this work, we outline a production asynchronous FL system design. Our work tackles the aforementioned issues, sketches of some of the system design challenges and their solutions, and touches upon principles that emerged from building a production FL system for millions of clients. Empirically, we demonstrate that asynchronous FL converges faster than synchronous FL when training across nearly one hundred million devices. In particular, in high concurrency settings, asynchronous FL is 5x faster and has nearly 8x less communication overhead than synchronous FL.

READ FULL TEXT
research
06/11/2021

Federated Learning with Buffered Asynchronous Aggregation

Federated Learning (FL) trains a shared model across distributed devices...
research
10/12/2020

FedAT: A Communication-Efficient Federated Learning Method with Asynchronous Tiers under Non-IID Data

Federated learning (FL) involves training a model over massive distribut...
research
12/15/2021

Blockchain-enabled Server-less Federated Learning

Motivated by the heterogeneous nature of devices participating in large-...
research
06/26/2022

Cross-Silo Federated Learning: Challenges and Opportunities

Federated learning (FL) is an emerging technology that enables the train...
research
10/28/2022

Efficient and Light-Weight Federated Learning via Asynchronous Distributed Dropout

Asynchronous learning protocols have regained attention lately, especial...
research
08/25/2023

DAG-ACFL: Asynchronous Clustered Federated Learning based on DAG-DLT

Federated learning (FL) aims to collaboratively train a global model whi...
research
02/14/2023

Federated Learning via Indirect Server-Client Communications

Federated Learning (FL) is a communication-efficient and privacy-preserv...

Please sign up or login with your details

Forgot password? Click here to reset