Re-Weighted Softmax Cross-Entropy to Control Forgetting in Federated Learning

04/11/2023
by   Gwen Legate, et al.
0

In Federated Learning, a global model is learned by aggregating model updates computed at a set of independent client nodes, to reduce communication costs multiple gradient steps are performed at each node prior to aggregation. A key challenge in this setting is data heterogeneity across clients resulting in differing local objectives which can lead clients to overly minimize their own local objective, diverging from the global solution. We demonstrate that individual client models experience a catastrophic forgetting with respect to data from other clients and propose an efficient approach that modifies the cross-entropy objective on a per-client basis by re-weighting the softmax logits prior to computing the loss. This approach shields classes outside a client's label set from abrupt representation change and we empirically demonstrate it can alleviate client forgetting and provide consistent improvements to standard federated learning algorithms. Our method is particularly beneficial under the most challenging federated learning settings where data heterogeneity is high and client participation in each round is low.

READ FULL TEXT

page 6

page 13

research
03/09/2022

Efficient Image Representation Learning with Federated Sampled Softmax

Learning image representations on decentralized data can bring many bene...
research
01/01/2023

Federated Learning with Client-Exclusive Classes

Existing federated classification algorithms typically assume the local ...
research
07/25/2023

FedDRL: A Trustworthy Federated Learning Model Fusion Method Based on Staged Reinforcement Learning

Traditional federated learning uses the number of samples to calculate t...
research
07/10/2023

FedYolo: Augmenting Federated Learning with Pretrained Transformers

The growth and diversity of machine learning applications motivate a ret...
research
03/11/2023

Stabilizing and Improving Federated Learning with Non-IID Data and Client Dropout

The label distribution skew induced data heterogeniety has been shown to...
research
09/01/2022

Federated Learning with Label Distribution Skew via Logits Calibration

Traditional federated optimization methods perform poorly with heterogen...
research
06/16/2023

HePCo: Data-Free Heterogeneous Prompt Consolidation for Continual Federated Learning

In this paper, we focus on the important yet understudied problem of Con...

Please sign up or login with your details

Forgot password? Click here to reset