FedLALR: Client-Specific Adaptive Learning Rates Achieve Linear Speedup for Non-IID Data

09/18/2023
by   Hao Sun, et al.
0

Federated learning is an emerging distributed machine learning method, enables a large number of clients to train a model without exchanging their local data. The time cost of communication is an essential bottleneck in federated learning, especially for training large-scale deep neural networks. Some communication-efficient federated learning methods, such as FedAvg and FedAdam, share the same learning rate across different clients. But they are not efficient when data is heterogeneous. To maximize the performance of optimization methods, the main challenge is how to adjust the learning rate without hurting the convergence. In this paper, we propose a heterogeneous local variant of AMSGrad, named FedLALR, in which each client adjusts its learning rate based on local historical gradient squares and synchronized learning rates. Theoretical analysis shows that our client-specified auto-tuned learning rate scheduling can converge and achieve linear speedup with respect to the number of clients, which enables promising scalability in federated optimization. We also empirically compare our method with several communication-efficient federated optimization methods. Extensive experimental results on Computer Vision (CV) tasks and Natural Language Processing (NLP) task show the efficacy of our proposed FedLALR method and also coincides with our theoretical findings.

READ FULL TEXT

page 1

page 5

research
02/29/2020

Adaptive Federated Optimization

Federated learning is a distributed machine learning paradigm in which a...
research
02/10/2023

Achieving Linear Speedup in Non-IID Federated Bilevel Learning

Federated bilevel optimization has received increasing attention in vari...
research
02/18/2020

Distributed Non-Convex Optimization with Sublinear Speedup under Intermittent Client Availability

Federated learning is a new distributed machine learning framework, wher...
research
01/25/2023

When to Trust Aggregated Gradients: Addressing Negative Client Sampling in Federated Learning

Federated Learning has become a widely-used framework which allows learn...
research
03/27/2023

Adaptive Federated Learning via New Entropy Approach

Federated Learning (FL) has recently emerged as a popular framework, whi...
research
09/14/2020

Effective Federated Adaptive Gradient Methods with Non-IID Decentralized Data

Federated learning allows loads of edge computing devices to collaborati...
research
08/08/2023

Federated Zeroth-Order Optimization using Trajectory-Informed Surrogate Gradients

Federated optimization, an emerging paradigm which finds wide real-world...

Please sign up or login with your details

Forgot password? Click here to reset