Faster Federated Learning with Decaying Number of Local SGD Steps

05/16/2023
by   Jed Mills, et al.
0

In Federated Learning (FL) client devices connected over the internet collaboratively train a machine learning model without sharing their private data with a central server or with other clients. The seminal Federated Averaging (FedAvg) algorithm trains a single global model by performing rounds of local training on clients followed by model averaging. FedAvg can improve the communication-efficiency of training by performing more steps of Stochastic Gradient Descent (SGD) on clients in each round. However, client data in real-world FL is highly heterogeneous, which has been extensively shown to slow model convergence and harm final performance when K > 1 steps of SGD are performed on clients per round. In this work we propose decaying K as training progresses, which can jointly improve the final performance of the FL model whilst reducing the wall-clock time and the total computational cost of training compared to using a fixed K. We analyse the convergence of FedAvg with decaying K for strongly-convex objectives, providing novel insights into the convergence properties, and derive three theoretically-motivated decay schedules for K. We then perform thorough experiments on four benchmark FL datasets (FEMNIST, CIFAR100, Sentiment140, Shakespeare) to show the real-world benefit of our approaches in terms of real-world convergence time, computational cost, and generalisation performance.

READ FULL TEXT

page 7

page 8

research
02/12/2022

On Federated Learning with Energy Harvesting Clients

Catering to the proliferation of Internet of Things devices and distribu...
research
02/20/2022

Personalized Federated Learning with Exact Stochastic Gradient Descent

In Federated Learning (FL), datasets across clients tend to be heterogen...
research
02/12/2020

Federated Clustering via Matrix Factorization Models: From Model Averaging to Gradient Sharing

Recently, federated learning (FL) has drawn significant attention due to...
research
08/20/2021

Accelerating Federated Learning with a Global Biased Optimiser

Federated Learning (FL) is a recent development in the field of machine ...
research
09/25/2022

On the Stability Analysis of Open Federated Learning Systems

We consider the open federated learning (FL) systems, where clients may ...
research
10/08/2021

FSL: Federated Supermask Learning

Federated learning (FL) allows multiple clients with (private) data to c...
research
04/26/2022

Federated Stochastic Primal-dual Learning with Differential Privacy

Federated learning (FL) is a new paradigm that enables many clients to j...

Please sign up or login with your details

Forgot password? Click here to reset