Federated learning with incremental clustering for heterogeneous data

Federated learning enables different parties to collaboratively build a global model under the orchestration of a server while keeping the training data on clients' devices. However, performance is affected when clients have heterogeneous data. To cope with this problem, we assume that despite data heterogeneity, there are groups of clients who have similar data distributions that can be clustered. In previous approaches, in order to cluster clients the server requires clients to send their parameters simultaneously. However, this can be problematic in a context where there is a significant number of participants that may have limited availability. To prevent such a bottleneck, we propose FLIC (Federated Learning with Incremental Clustering), in which the server exploits the updates sent by clients during federated training instead of asking them to send their parameters simultaneously. Hence no additional communications between the server and the clients are necessary other than what classical federated learning requires. We empirically demonstrate for various non-IID cases that our approach successfully splits clients into groups following the same data distributions. We also identify the limitations of FLIC by studying its capability to partition clients at the early stages of the federated learning process efficiently. We further address attacks on models as a form of data heterogeneity and empirically show that FLIC is a robust defense against poisoning attacks even when the proportion of malicious clients is higher than 50%.

READ FULL TEXT
research
03/28/2020

Federated Residual Learning

We study a new form of federated learning where the clients train person...
research
03/22/2023

Prototype Helps Federated Learning: Towards Faster Convergence

Federated learning (FL) is a distributed machine learning technique in w...
research
05/23/2022

Orchestra: Unsupervised Federated Learning via Globally Consistent Clustering

Federated learning is generally used in tasks where labels are readily a...
research
07/18/2021

Federated Action Recognition on Heterogeneous Embedded Devices

Federated learning allows a large number of devices to jointly learn a m...
research
06/15/2023

Towards Practical Federated Causal Structure Learning

Understanding causal relations is vital in scientific discovery. The pro...
research
08/11/2020

Federated Learning via Synthetic Data

Federated learning allows for the training of a model using data on mult...
research
04/25/2023

Mobilizing Personalized Federated Learning via Random Walk Stochastic ADMM

In this research, we investigate the barriers associated with implementi...

Please sign up or login with your details

Forgot password? Click here to reset