Rate Region for Indirect Multiterminal Source Coding in Federated Learning

01/21/2021
by   Naifu Zhang, et al.
5

One of the main focus in federated learning (FL) is the communication efficiency since a large number of participating edge devices send their updates to the edge server at each round of the model training. Existing works reconstruct each model update from edge devices and implicitly assume that the local model updates are independent over edge device. In FL, however, the model update is an indirect multi-terminal source coding problem where each edge device cannot observe directly the source that is to be reconstructed at the decoder, but is rather provided only with a noisy version. The existing works do not leverage the redundancy in the information transmitted by different edges. This paper studies the rate region for the indirect multiterminal source coding problem in FL. The goal is to obtain the minimum achievable rate at a particular upper bound of gradient variance. We obtain the rate region for multiple edge devices in general case and derive an explicit formula of the sum-rate distortion function in the special case where gradient are identical over edge device and dimension. Finally, we analysis communication efficiency of convex Mini-batched SGD and non-convex Minibatched SGD based on the sum-rate distortion function, respectively.

READ FULL TEXT

page 5

page 6

page 7

page 8

page 10

page 27

page 28

page 30

research
09/18/2021

Toward Efficient Federated Learning in Multi-Channeled Mobile Edge Network with Layerd Gradient Compression

A fundamental issue for federated learning (FL) is how to achieve optima...
research
01/28/2020

Update Aware Device Scheduling for Federated Learning at the Wireless Edge

We study federated learning (FL) at the wireless edge, where power-limit...
research
06/02/2022

Resource Allocation for Compression-aided Federated Learning with High Distortion Rate

Recently, a considerable amount of works have been made to tackle the co...
research
03/03/2020

Adaptive Federated Learning With Gradient Compression in Uplink NOMA

Federated learning (FL) is an emerging machine learning technique that a...
research
01/24/2022

Communication-Efficient Stochastic Zeroth-Order Optimization for Federated Learning

Federated learning (FL), as an emerging edge artificial intelligence par...
research
06/28/2022

Fundamental Limits of Communication Efficiency for Model Aggregation in Distributed Learning: A Rate-Distortion Approach

One of the main focuses in distributed learning is communication efficie...
research
10/25/2021

Optimization-Based GenQSGD for Federated Edge Learning

Optimal algorithm design for federated learning (FL) remains an open pro...

Please sign up or login with your details

Forgot password? Click here to reset