Taming Fat-Tailed ("Heavier-Tailed” with Potentially Infinite Variance) Noise in Federated Learning

10/03/2022
by   Haibo Yang, et al.
0

A key assumption in most existing works on FL algorithms' convergence analysis is that the noise in stochastic first-order information has a finite variance. Although this assumption covers all light-tailed (i.e., sub-exponential) and some heavy-tailed noise distributions (e.g., log-normal, Weibull, and some Pareto distributions), it fails for many fat-tailed noise distributions (i.e., “heavier-tailed” with potentially infinite variance) that have been empirically observed in the FL literature. To date, it remains unclear whether one can design convergent algorithms for FL systems that experience fat-tailed noise. This motivates us to fill this gap in this paper by proposing an algorithmic framework called FAT-Clipping (federated averaging with two-sided learning rates and clipping), which contains two variants: FAT-Clipping per-round (FAT-Clipping-PR) and FAT-Clipping per-iteration (FAT-Clipping-PI). Specifically, for the largest α∈ (1,2] such that the fat-tailed noise in FL still has a bounded α-moment, we show that both variants achieve 𝒪((mT)^2-α/α) and 𝒪((mT)^1-α/3α-2) convergence rates in the strongly-convex and general non-convex settings, respectively, where m and T are the numbers of clients and communication rounds. Moreover, at the expense of more clipping operations compared to FAT-Clipping-PR, FAT-Clipping-PI further enjoys a linear speedup effect with respect to the number of local updates at each client and being lower-bound-matching (i.e., order-optimal). Collectively, our results advance the understanding of designing efficient algorithms for FL systems that exhibit fat-tailed first-order oracle information.

READ FULL TEXT
research
07/11/2020

Federated Learning's Blessing: FedAvg has Linear Speedup

Federated learning (FL) learns a model jointly from a set of participati...
research
05/26/2022

A Unified Analysis of Federated Learning with Arbitrary Client Participation

Federated learning (FL) faces challenges of intermittent client availabi...
research
09/25/2022

On the Stability Analysis of Open Federated Learning Systems

We consider the open federated learning (FL) systems, where clients may ...
research
06/30/2022

Towards Federated Long-Tailed Learning

Data privacy and class imbalance are the norm rather than the exception ...
research
10/25/2022

SWIFT: Rapid Decentralized Federated Learning via Wait-Free Model Communication

The decentralized Federated Learning (FL) setting avoids the role of a p...
research
07/14/2023

Improved Convergence Analysis and SNR Control Strategies for Federated Learning in the Presence of Noise

We propose an improved convergence analysis technique that characterizes...
research
03/22/2023

Stochastic Nonsmooth Convex Optimization with Heavy-Tailed Noises

Recently, several studies consider the stochastic optimization problem b...

Please sign up or login with your details

Forgot password? Click here to reset