Improving the Transient Times for Distributed Stochastic Gradient Methods

05/11/2021
by   Kun Huang, et al.
0

We consider the distributed optimization problem where n agents each possessing a local cost function, collaboratively minimize the average of the n cost functions over a connected network. Assuming stochastic gradient information is available, we study a distributed stochastic gradient algorithm, called exact diffusion with adaptive stepsizes (EDAS) adapted from the Exact Diffusion method and NIDS and perform a non-asymptotic convergence analysis. We not only show that EDAS asymptotically achieves the same network independent convergence rate as centralized stochastic gradient descent (SGD) for minimizing strongly convex and smooth objective functions, but also characterize the transient time needed for the algorithm to approach the asymptotic convergence rate, which behaves as K_T=𝒪(n/1-λ_2), where 1-λ_2 stands for the spectral gap of the mixing matrix. To the best of our knowledge, EDAS achieves the shortest transient time when the average of the n cost functions is strongly convex and each cost function is smooth. Numerical simulations further corroborate and strengthen the obtained theoretical results.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/06/2019

A Non-Asymptotic Analysis of Network Independence for Distributed Stochastic Gradient Descent

This paper is concerned with minimizing the average of n cost functions ...
research
01/14/2023

CEDAS: A Compressed Decentralized Stochastic Gradient Method with Improved Convergence

In this paper, we consider solving the distributed optimization problem ...
research
05/17/2021

Removing Data Heterogeneity Influence Enhances Network Topology Dependence of Decentralized SGD

We consider decentralized stochastic optimization problems where a netwo...
research
10/28/2022

Secure Distributed Optimization Under Gradient Attacks

In this paper, we study secure distributed optimization against arbitrar...
research
05/26/2021

Distributed Zeroth-Order Stochastic Optimization in Time-varying Networks

We consider a distributed convex optimization problem in a network which...
research
03/04/2022

Analysis of closed-loop inertial gradient dynamics

In this paper, we analyse the performance of the closed-loop Whiplash gr...
research
06/21/2023

Distributed Random Reshuffling Methods with Improved Convergence

This paper proposes two distributed random reshuffling methods, namely G...

Please sign up or login with your details

Forgot password? Click here to reset