Federated Learning Using Three-Operator ADMM

11/08/2022
by   Shashi Kant, et al.
0

Federated learning (FL) has emerged as an instance of distributed machine learning paradigm that avoids the transmission of data generated on the users' side. Although data are not transmitted, edge devices have to deal with limited communication bandwidths, data heterogeneity, and straggler effects due to the limited computational resources of users' devices. A prominent approach to overcome such difficulties is FedADMM, which is based on the classical two-operator consensus alternating direction method of multipliers (ADMM). The common assumption of FL algorithms, including FedADMM, is that they learn a global model using data only on the users' side and not on the edge server. However, in edge learning, the server is expected to be near the base station and have direct access to rich datasets. In this paper, we argue that leveraging the rich data on the edge server is much more beneficial than utilizing only user datasets. Specifically, we show that the mere application of FL with an additional virtual user node representing the data on the edge server is inefficient. We propose FedTOP-ADMM, which generalizes FedADMM and is based on a three-operator ADMM-type technique that exploits a smooth cost function on the edge server to learn a global model parallel to the edge devices. Our numerical experiments indicate that FedTOP-ADMM has substantial gain up to 33% in communication efficiency to reach a desired test accuracy with respect to FedADMM, including a virtual user on the edge server.

READ FULL TEXT

page 14

page 15

page 18

research
05/30/2022

Confederated Learning: Federated Learning with Decentralized Edge Servers

Federated learning (FL) is an emerging machine learning paradigm that al...
research
08/16/2022

Resource-aware Federated Learning using Knowledge Extraction and Multi-model Fusion

With increasing concern about user data privacy, federated learning (FL)...
research
09/05/2019

Hierarchical Federated Learning Across Heterogeneous Cellular Networks

We study collaborative machine learning (ML) across wireless devices, ea...
research
10/29/2022

Fast-Convergent Federated Learning via Cyclic Aggregation

Federated learning (FL) aims at optimizing a shared global model over mu...
research
11/02/2021

FedFly: Towards Migration in Edge-based Distributed Federated Learning

Federated learning (FL) is a privacy-preserving distributed machine lear...
research
04/25/2022

FedDUAP: Federated Learning with Dynamic Update and Adaptive Pruning Using Shared Data on the Server

Despite achieving remarkable performance, Federated Learning (FL) suffer...
research
03/15/2022

SemiPFL: Personalized Semi-Supervised Federated Learning Framework for Edge Intelligence

Recent advances in wearable devices and Internet-of-Things (IoT) have le...

Please sign up or login with your details

Forgot password? Click here to reset