Communication-Efficient Federated Learning through Importance Sampling

06/22/2023
by   Berivan Isik, et al.
0

The high communication cost of sending model updates from the clients to the server is a significant bottleneck for scalable federated learning (FL). Among existing approaches, state-of-the-art bitrate-accuracy tradeoffs have been achieved using stochastic compression methods – in which the client n sends a sample from a client-only probability distribution q_ϕ^(n), and the server estimates the mean of the clients' distributions using these samples. However, such methods do not take full advantage of the FL setup where the server, throughout the training process, has side information in the form of a pre-data distribution p_θ that is close to the client's distribution q_ϕ^(n) in Kullback-Leibler (KL) divergence. In this work, we exploit this closeness between the clients' distributions q_ϕ^(n)'s and the side information p_θ at the server, and propose a framework that requires approximately D_KL(q_ϕ^(n)|| p_θ) bits of communication. We show that our method can be integrated into many existing stochastic compression frameworks such as FedPM, Federated SGLD, and QSGD to attain the same (and often higher) test accuracy with up to 50 times reduction in the bitrate.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
11/17/2022

Improving Federated Learning Communication Efficiency with Global Momentum Fusion for Gradient Compression Schemes

Communication costs within Federated learning hinder the system scalabil...
research
07/30/2020

Communication-Efficient Federated Learning via Optimal Client Sampling

Federated learning is a private and efficient framework for learning mod...
research
07/28/2022

FedVARP: Tackling the Variance Due to Partial Client Participation in Federated Learning

Data-heterogeneous federated learning (FL) systems suffer from two signi...
research
04/04/2023

Privacy Amplification via Compression: Achieving the Optimal Privacy-Accuracy-Communication Trade-off in Distributed Mean Estimation

Privacy and communication constraints are two major bottlenecks in feder...
research
12/11/2021

Server-Side Local Gradient Averaging and Learning Rate Acceleration for Scalable Split Learning

In recent years, there have been great advances in the field of decentra...
research
08/17/2020

Shuffled Model of Federated Learning: Privacy, Communication and Accuracy Trade-offs

We consider a distributed empirical risk minimization (ERM) optimization...
research
11/14/2020

CatFedAvg: Optimising Communication-efficiency and Classification Accuracy in Federated Learning

Federated learning has allowed the training of statistical models over r...

Please sign up or login with your details

Forgot password? Click here to reset