Accelerating Federated Learning via Sampling Anchor Clients with Large Batches

06/13/2022
by   Feijie Wu, et al.
0

Using large batches in recent federated learning studies has improved convergence rates, but it requires additional computation overhead compared to using small batches. To overcome this limitation, we propose a unified framework FedAMD, which disjoints the participants into anchor and miner groups based on time-varying probabilities. Each client in the anchor group computes the gradient using a large batch, which is regarded as its bullseye. Clients in the miner group perform multiple local updates using serial mini-batches, and each local update is also indirectly regulated by the global target derived from the average of clients' bullseyes. As a result, the miner group follows a near-optimal update towards the global minimizer, adapted to update the global model. Measured by ϵ-approximation, FedAMD achieves a convergence rate of O(1/ϵ) under non-convex objectives by sampling an anchor with a constant probability. The theoretical result considerably surpasses the state-of-the-art algorithm BVR-L-SGD at O(1/ϵ^3/2), while FedAMD reduces at least O(1/ϵ) communication overhead. Empirical studies on real-world datasets validate the effectiveness of FedAMD and demonstrate the superiority of our proposed algorithm.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/23/2022

Semi-Decentralized Federated Learning with Collaborative Relaying

We present a semi-decentralized federated learning algorithm wherein cli...
research
01/10/2022

Communication-Efficient Federated Learning with Acceleration of Global Momentum

Federated learning often suffers from unstable and slow convergence due ...
research
05/13/2022

Federated Learning Under Intermittent Client Availability and Time-Varying Communication Constraints

Federated learning systems facilitate training of global models in setti...
research
10/26/2020

Optimal Client Sampling for Federated Learning

It is well understood that client-master communication can be a primary ...
research
01/26/2022

Server-Side Stepsizes and Sampling Without Replacement Provably Help in Federated Optimization

We present a theoretical study of server-side optimization in federated ...
research
09/16/2021

Federated Submodel Averaging

We study practical data characteristics underlying federated learning, w...
research
06/25/2021

Implicit Gradient Alignment in Distributed and Federated Learning

A major obstacle to achieving global convergence in distributed and fede...

Please sign up or login with your details

Forgot password? Click here to reset