When to Trust Aggregated Gradients: Addressing Negative Client Sampling in Federated Learning

01/25/2023
by   Wenkai Yang, et al.
0

Federated Learning has become a widely-used framework which allows learning a global model on decentralized local datasets under the condition of protecting local data privacy. However, federated learning faces severe optimization difficulty when training samples are not independently and identically distributed (non-i.i.d.). In this paper, we point out that the client sampling practice plays a decisive role in the aforementioned optimization difficulty. We find that the negative client sampling will cause the merged data distribution of currently sampled clients heavily inconsistent with that of all available clients, and further make the aggregated gradient unreliable. To address this issue, we propose a novel learning rate adaptation mechanism to adaptively adjust the server learning rate for the aggregated gradient in each round, according to the consistency between the merged data distribution of currently sampled clients and that of all available clients. Specifically, we make theoretical deductions to find a meaningful and robust indicator that is positively related to the optimal server learning rate and can effectively reflect the merged data distribution of sampled clients, and we utilize it for the server learning rate adaptation. Extensive experiments on multiple image and text classification tasks validate the great effectiveness of our method.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
07/09/2020

Client Adaptation improves Federated Learning with Simulated Non-IID Clients

We present a federated learning approach for learning a client adaptable...
research
03/28/2023

Fast Convergent Federated Learning with Aggregated Gradients

Federated Learning (FL) is a novel machine learning framework, which ena...
research
09/18/2023

FedLALR: Client-Specific Adaptive Learning Rates Achieve Linear Speedup for Non-IID Data

Federated learning is an emerging distributed machine learning method, e...
research
11/25/2022

Federated Graph-based Sampling with Arbitrary Client Availability

While federated learning has shown strong results in optimizing a machin...
research
03/27/2023

Adaptive Federated Learning via New Entropy Approach

Federated Learning (FL) has recently emerged as a popular framework, whi...
research
10/22/2021

PPSGCN: A Privacy-Preserving Subgraph Sampling Based Distributed GCN Training Method

Graph convolutional networks (GCNs) have been widely adopted for graph r...
research
06/06/2022

Rate-Distortion Theoretic Bounds on Generalization Error for Distributed Learning

In this paper, we use tools from rate-distortion theory to establish new...

Please sign up or login with your details

Forgot password? Click here to reset