Unlocking the Potential of Federated Learning for Deeper Models

06/05/2023
by   Haolin Wang, et al.
0

Federated learning (FL) is a new paradigm for distributed machine learning that allows a global model to be trained across multiple clients without compromising their privacy. Although FL has demonstrated remarkable success in various scenarios, recent studies mainly utilize shallow and small neural networks. In our research, we discover a significant performance decline when applying the existing FL framework to deeper neural networks, even when client data are independently and identically distributed (i.i.d.). Our further investigation shows that the decline is due to the continuous accumulation of dissimilarities among client models during the layer-by-layer back-propagation process, which we refer to as "divergence accumulation." As deeper models involve a longer chain of divergence accumulation, they tend to manifest greater divergence, subsequently leading to performance decline. Both theoretical derivations and empirical evidence are proposed to support the existence of divergence accumulation and its amplified effects in deeper models. To address this issue, we propose several technical guidelines based on reducing divergence, such as using wider models and reducing the receptive field. These approaches can greatly improve the accuracy of FL on deeper models. For example, the application of these guidelines can boost the ResNet101 model's performance by as much as 43% on the Tiny-ImageNet dataset.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
03/16/2022

Client-Wise Targeted Backdoor in Federated Learning

Federated Learning (FL) emerges from the privacy concerns traditional ma...
research
11/23/2020

LINDT: Tackling Negative Federated Learning with Local Adaptation

Federated Learning (FL) is a promising distributed learning paradigm, wh...
research
10/04/2019

Clustered Federated Learning: Model-Agnostic Distributed Multi-Task Optimization under Privacy Constraints

Federated Learning (FL) is currently the most widely adopted framework f...
research
11/07/2022

Closing the Gap between Client and Global Model Performance in Heterogeneous Federated Learning

The heterogeneity of hardware and data is a well-known and studied probl...
research
02/15/2022

Architecture Agnostic Federated Learning for Neural Networks

With growing concerns regarding data privacy and rapid increase in data ...
research
07/14/2023

L-DAWA: Layer-wise Divergence Aware Weight Aggregation in Federated Self-Supervised Visual Representation Learning

The ubiquity of camera-enabled devices has led to large amounts of unlab...

Please sign up or login with your details

Forgot password? Click here to reset