Implicit Gradient Alignment in Distributed and Federated Learning

06/25/2021
by   Yatin Dandi, et al.
12

A major obstacle to achieving global convergence in distributed and federated learning is the misalignment of gradients across clients, or mini-batches due to heterogeneity and stochasticity of the distributed data. One way to alleviate this problem is to encourage the alignment of gradients across different clients throughout training. Our analysis reveals that this goal can be accomplished by utilizing the right optimization method that replicates the implicit regularization effect of SGD, leading to gradient alignment as well as improvements in test accuracies. Since the existence of this regularization in SGD completely relies on the sequential use of different mini-batches during training, it is inherently absent when training with large mini-batches. To obtain the generalization benefits of this regularization while increasing parallelism, we propose a novel GradAlign algorithm that induces the same implicit regularization while allowing the use of arbitrarily large batches in each update. We experimentally validate the benefit of our algorithm in different distributed and federated learning settings.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
08/19/2021

Communication-Efficient Federated Learning via Robust Distributed Mean Estimation

Federated learning commonly relies on algorithms such as distributed (mi...
research
07/20/2023

Boosting Federated Learning Convergence with Prototype Regularization

As a distributed machine learning technique, federated learning (FL) req...
research
08/01/2021

A Decentralized Federated Learning Framework via Committee Mechanism with Convergence Guarantee

Federated learning allows multiple participants to collaboratively train...
research
06/01/2022

Optimization with access to auxiliary information

We investigate the fundamental optimization question of minimizing a tar...
research
07/10/2022

FedSS: Federated Learning with Smart Selection of clients

Federated learning provides the ability to learn over heterogeneous user...
research
06/13/2022

Accelerating Federated Learning via Sampling Anchor Clients with Large Batches

Using large batches in recent federated learning studies has improved co...

Please sign up or login with your details

Forgot password? Click here to reset