Local Stochastic Gradient Descent Ascent: Convergence Analysis and Communication Efficiency

02/25/2021
by   Yuyang Deng, et al.
13

Local SGD is a promising approach to overcome the communication overhead in distributed learning by reducing the synchronization frequency among worker nodes. Despite the recent theoretical advances of local SGD in empirical risk minimization, the efficiency of its counterpart in minimax optimization remains unexplored. Motivated by large scale minimax learning problems, such as adversarial robust learning and training generative adversarial networks (GANs), we propose local Stochastic Gradient Descent Ascent (local SGDA), where the primal and dual variables can be trained locally and averaged periodically to significantly reduce the number of communications. We show that local SGDA can provably optimize distributed minimax problems in both homogeneous and heterogeneous data with reduced number of communications and establish convergence rates under strongly-convex-strongly-concave and nonconvex-strongly-concave settings. In addition, we propose a novel variant local SGDA+, to solve nonconvex-nonconcave problems. We give corroborating empirical evidence on different distributed minimax problems.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/17/2022

Tight Analysis of Extra-gradient and Optimistic Gradient Methods For Nonconvex Minimax Problems

Despite the established convergence theory of Optimistic Gradient Descen...
research
06/18/2021

Local AdaGrad-Type Algorithm for Stochastic Convex-Concave Minimax Problems

Large scale convex-concave minimax problems arise in numerous applicatio...
research
11/25/2021

Randomized Stochastic Gradient Descent Ascent

An increasing number of machine learning problems, such as robust or adv...
research
10/25/2020

Local SGD for Saddle-Point Problems

GAN is one of the most popular and commonly used neural network models. ...
research
06/02/2022

A Communication-efficient Algorithm with Linear Convergence for Federated Minimax Learning

In this paper, we study a large-scale multi-agent minimax optimization p...
research
06/09/2022

What is a Good Metric to Study Generalization of Minimax Learners?

Minimax optimization has served as the backbone of many machine learning...
research
05/26/2019

ODE Analysis of Stochastic Gradient Methods with Optimism and Anchoring for Minimax Problems and GANs

Despite remarkable empirical success, the training dynamics of generativ...

Please sign up or login with your details

Forgot password? Click here to reset