Federated Dynamic Sparse Training: Computing Less, Communicating Less, Yet Learning Better

12/18/2021
by   Sameer Bibikar, et al.
0

Federated learning (FL) enables distribution of machine learning workloads from the cloud to resource-limited edge devices. Unfortunately, current deep networks remain not only too compute-heavy for inference and training on edge devices, but also too large for communicating updates over bandwidth-constrained networks. In this paper, we develop, implement, and experimentally validate a novel FL framework termed Federated Dynamic Sparse Training (FedDST) by which complex neural networks can be deployed and trained with substantially improved efficiency in both on-device computation and in-network communication. At the core of FedDST is a dynamic process that extracts and trains sparse sub-networks from the target full network. With this scheme, "two birds are killed with one stone:" instead of full models, each client performs efficient training of its own sparse networks, and only sparse networks are transmitted between devices and the cloud. Furthermore, our results reveal that the dynamic sparsity during FL training more flexibly accommodates local heterogeneity in FL agents than the fixed, shared sparse masks. Moreover, dynamic sparsity naturally introduces an "in-time self-ensembling effect" into the training dynamics and improves the FL performance even over dense training. In a realistic and challenging non i.i.d. FL setting, FedDST consistently outperforms competing algorithms in our experiments: for instance, at any fixed upload data cap on non-iid CIFAR-10, it gains an impressive accuracy advantage of 10 upload data cap; the accuracy gap remains 3 upload data cap, further demonstrating efficacy of FedDST. Code is available at: https://github.com/bibikar/feddst.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
11/08/2022

Centaur: Federated Learning for Constrained Edge Devices

Federated learning (FL) on deep neural networks facilitates new applicat...
research
04/15/2023

SalientGrads: Sparse Models for Communication Efficient and Data Aware Distributed Federated Training

Federated learning (FL) enables the training of a model leveraging decen...
research
02/17/2022

FLAME: Federated Learning Across Multi-device Environments

Federated Learning (FL) enables distributed training of machine learning...
research
11/02/2021

FedFly: Towards Migration in Edge-based Distributed Federated Learning

Federated learning (FL) is a privacy-preserving distributed machine lear...
research
09/02/2023

Equitable-FL: Federated Learning with Sparsity for Resource-Constrained Environment

In Federated Learning, model training is performed across multiple compu...
research
09/13/2023

FedDIP: Federated Learning with Extreme Dynamic Pruning and Incremental Regularization

Federated Learning (FL) has been successfully adopted for distributed tr...
research
08/04/2022

ZeroFL: Efficient On-Device Training for Federated Learning with Local Sparsity

When the available hardware cannot meet the memory and compute requireme...

Please sign up or login with your details

Forgot password? Click here to reset