A Distributed Flexible Delay-tolerant Proximal Gradient Algorithm

06/25/2018
by   Konstantin Mishchenko, et al.
0

We develop and analyze an asynchronous algorithm for distributed convex optimization when the objective writes a sum of smooth functions, local to each worker, and a non-smooth function. Unlike many existing methods, our distributed algorithm is adjustable to various levels of communication cost, delays, machines computational power, and functions smoothness. A unique feature is that the stepsizes do not depend on communication delays nor number of machines, which is highly desirable for scalability. We prove that the algorithm converges linearly in the strongly convex case, and provide guarantees of convergence for the non-strongly convex case. The obtained rates are the same as the vanilla proximal gradient algorithm over some introduced epoch sequence that subsumes the delays of the system. We provide numerical results on large-scale machine learning problems to demonstrate the merits of the proposed method.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
01/28/2019

Asynchronous Accelerated Proximal Stochastic Gradient for Strongly Convex Distributed Finite Sums

In this work, we study the problem of minimizing the sum of strongly con...
research
06/23/2021

Learning Under Delayed Feedback: Implicitly Adapting to Gradient Delays

We consider stochastic convex optimization problems, where several machi...
research
05/20/2019

A Linearly Convergent Proximal Gradient Algorithm for Decentralized Optimization

Decentralized optimization is a promising paradigm that finds various ap...
research
06/05/2015

Communication Complexity of Distributed Convex Learning and Optimization

We study the fundamental limits to communication-efficient distributed m...
research
08/13/2018

AsySPA: An Exact Asynchronous Algorithm for Convex Optimization Over Digraphs

This paper proposes a novel exact asynchronous subgradient-push algorith...
research
07/21/2023

Robust Fully-Asynchronous Methods for Distributed Training over General Architecture

Perfect synchronization in distributed machine learning problems is inef...

Please sign up or login with your details

Forgot password? Click here to reset