Experimenting with Normalization Layers in Federated Learning on non-IID scenarios

03/19/2023
by   Bruno Casella, et al.
0

Training Deep Learning (DL) models require large, high-quality datasets, often assembled with data from different institutions. Federated Learning (FL) has been emerging as a method for privacy-preserving pooling of datasets employing collaborative training from different institutions by iteratively globally aggregating locally trained models. One critical performance challenge of FL is operating on datasets not independently and identically distributed (non-IID) among the federation participants. Even though this fragility cannot be eliminated, it can be debunked by a suitable optimization of two hyper-parameters: layer normalization methods and collaboration frequency selection. In this work, we benchmark five different normalization layers for training Neural Networks (NNs), two families of non-IID data skew, and two datasets. Results show that Batch Normalization, widely employed for centralized DL, is not the best choice for FL, whereas Group and Layer Normalization consistently outperform Batch Normalization. Similarly, frequent model aggregation decreases convergence speed and mode quality.

READ FULL TEXT
research
03/12/2023

Making Batch Normalization Great in Federated Deep Learning

Batch Normalization (BN) is commonly used in modern deep neural networks...
research
01/08/2023

Why Batch Normalization Damage Federated Learning on Non-IID Data?

As a promising distributed learning paradigm, federated learning (FL) in...
research
03/10/2023

Optimizing Federated Learning for Medical Image Classification on Distributed Non-iid Datasets with Partial Labels

Numerous large-scale chest x-ray datasets have spearheaded expert-level ...
research
08/18/2023

Normalization Is All You Need: Understanding Layer-Normalized Federated Learning under Extreme Label Shift

Layer normalization (LN) is a widely adopted deep learning technique esp...
research
04/02/2023

FedFTN: Personalized Federated Learning with Deep Feature Transformation Network for Multi-institutional Low-count PET Denoising

Low-count PET is an efficient way to reduce radiation exposure and acqui...
research
03/31/2023

Benchmarking FedAvg and FedCurv for Image Classification Tasks

Classic Machine Learning techniques require training on data available i...
research
04/15/2021

See through Gradients: Image Batch Recovery via GradInversion

Training deep neural networks requires gradient estimation from data bat...

Please sign up or login with your details

Forgot password? Click here to reset