Over-the-Air Federated Learning from Heterogeneous Data

09/27/2020
by   Tomer Sery, et al.
0

Federated learning (FL) is a framework for distributed learning of centralized models. In FL, a set of edge devices train a model using their local data, while repeatedly exchanging their trained updates with a central server. This procedure allows tuning a centralized model in a distributed fashion without having the users share their possibly private data. In this paper, we focus on over-the-air (OTA) FL, which has been suggested recently to reduce the communication overhead of FL due to the repeated transmissions of the model updates by a large number of users over the wireless channel. In OTA FL, all users simultaneously transmit their updates as analog signals over a multiple access channel, and the server receives a superposition of the analog transmitted signals. However, this approach results in the channel noise directly affecting the optimization procedure, which may degrade the accuracy of the trained model. We develop a Convergent OTA FL (COTAF) algorithm which enhances the common local stochastic gradient descent (SGD) FL algorithm, introducing precoding at the users and scaling at the server, which gradually mitigates the effect of the noise. We analyze the convergence of COTAF to the loss minimizing model and quantify the effect of a statistically heterogeneous setup, i.e. when the training data of each user obeys a different distribution. Our analysis reveals the ability of COTAF to achieve a convergence rate similar to that achievable over error-free channels. Our simulations demonstrate the improved convergence of COTAF over vanilla OTA local SGD for training using non-synthetic datasets. Furthermore, we numerically show that the precoding induced by COTAF notably improves the convergence rate and the accuracy of models trained via OTA FL.

READ FULL TEXT
research
05/12/2022

Over-the-Air Federated Learning with Joint Adaptive Computation and Power Control

This paper considers over-the-air federated learning (OTA-FL). OTA-FL ex...
research
11/17/2020

Private Wireless Federated Learning with Anonymous Over-the-Air Computation

We study the problem of differentially private wireless federated learni...
research
08/20/2019

On Analog Gradient Descent Learning over Multiple Access Fading Channels

We consider a distributed learning problem over multiple access channel ...
research
04/15/2022

Server Free Wireless Federated Learning: Architecture, Algorithm, and Analysis

We demonstrate that merely analog transmissions and match filtering can ...
research
06/05/2020

UVeQFed: Universal Vector Quantization for Federated Learning

Traditional deep learning models are trained at a centralized server usi...
research
10/27/2022

Decentralized Federated Learning via Non-Coherent Over-the-Air Consensus

This paper presents NCOTA-DGD, a Decentralized Gradient Descent (DGD) al...
research
07/14/2021

Communication-Efficient Hierarchical Federated Learning for IoT Heterogeneous Systems with Imbalanced Data

Federated learning (FL) is a distributed learning methodology that allow...

Please sign up or login with your details

Forgot password? Click here to reset