Is Normalization Indispensable for Multi-domain Federated Learning?

06/09/2023
by   Weiming Zhuang, et al.
0

Federated learning (FL) enhances data privacy with collaborative in-situ training on decentralized clients. Nevertheless, FL encounters challenges due to non-independent and identically distributed (non-i.i.d) data, leading to potential performance degradation and hindered convergence. While prior studies predominantly addressed the issue of skewed label distribution, our research addresses a crucial yet frequently overlooked problem known as multi-domain FL. In this scenario, clients' data originate from diverse domains with distinct feature distributions, as opposed to label distributions. To address the multi-domain problem in FL, we propose a novel method called Federated learning Without normalizations (FedWon). FedWon draws inspiration from the observation that batch normalization (BN) faces challenges in effectively modeling the statistics of multiple domains, while alternative normalization techniques possess their own limitations. In order to address these issues, FedWon eliminates all normalizations in FL and reparameterizes convolution layers with scaled weight standardization. Through comprehensive experimentation on four datasets and four models, our results demonstrate that FedWon surpasses both FedAvg and the current state-of-the-art method (FedBN) across all experimental setups, achieving notable improvements of over 10 Furthermore, FedWon is versatile for both cross-silo and cross-device FL, exhibiting strong performance even with a batch size as small as 1, thereby catering to resource-constrained devices. Additionally, FedWon effectively tackles the challenge of skewed label distribution.

READ FULL TEXT
research
06/10/2023

Optimizing the Collaboration Structure in Cross-Silo Federated Learning

In federated learning (FL), multiple clients collaborate to train machin...
research
01/08/2023

Why Batch Normalization Damage Federated Learning on Non-IID Data?

As a promising distributed learning paradigm, federated learning (FL) in...
research
08/18/2023

Normalization Is All You Need: Understanding Layer-Normalized Federated Learning under Extreme Label Shift

Layer normalization (LN) is a widely adopted deep learning technique esp...
research
03/12/2023

Making Batch Normalization Great in Federated Deep Learning

Batch Normalization (BN) is commonly used in modern deep neural networks...
research
06/18/2021

Federated Robustness Propagation: Sharing Adversarial Robustness in Federated Learning

Federated learning (FL) emerges as a popular distributed learning schema...
research
03/10/2023

Optimizing Federated Learning for Medical Image Classification on Distributed Non-iid Datasets with Partial Labels

Numerous large-scale chest x-ray datasets have spearheaded expert-level ...
research
07/26/2021

Aggregate or Not? Exploring Where to Privatize in DNN Based Federated Learning Under Different Non-IID Scenes

Although federated learning (FL) has recently been proposed for efficien...

Please sign up or login with your details

Forgot password? Click here to reset