DSCOVR: Randomized Primal-Dual Block Coordinate Algorithms for Asynchronous Distributed Optimization

10/13/2017
by   Lin Xiao, et al.
0

Machine learning with big data often involves large optimization models. For distributed optimization over a cluster of machines, frequent communication and synchronization of all model parameters (optimization variables) can be very costly. A promising solution is to use parameter servers to store different subsets of the model parameters, and update them asynchronously at different machines using local datasets. In this paper, we focus on distributed optimization of large linear models with convex loss functions, and propose a family of randomized primal-dual block coordinate algorithms that are especially suitable for asynchronous distributed implementation with parameter servers. In particular, we work with the saddle-point formulation of such problems which allows simultaneous data and model partitioning, and exploit its structure by doubly stochastic coordinate optimization with variance reduction (DSCOVR). Compared with other first-order distributed algorithms, we show that DSCOVR may require less amount of overall computation and communication, and less or no synchronization. We discuss the implementation details of the DSCOVR algorithms, and present numerical experiments on an industrial distributed computing system.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/18/2017

Asynchronous parallel primal-dual block update methods

Recent several years have witnessed the surge of asynchronous (async-) p...
research
03/24/2017

A randomized primal distributed algorithm for partitioned and big-data non-convex optimization

In this paper we consider a distributed optimization scenario in which t...
research
05/19/2018

Tell Me Something New: a new framework for asynchronous parallel learning

We present a novel approach for parallel computation in the context of m...
research
05/04/2021

The distributed dual ascent algorithm is robust to asynchrony

The distributed dual ascent is an established algorithm to solve strongl...
research
07/26/2021

Asynchronous Distributed Reinforcement Learning for LQR Control via Zeroth-Order Block Coordinate Descent

Recently introduced distributed zeroth-order optimization (ZOO) algorith...
research
10/31/2016

Optimization for Large-Scale Machine Learning with Distributed Features and Observations

As the size of modern data sets exceeds the disk and memory capacities o...
research
09/22/2014

Parallel and Distributed Block-Coordinate Frank-Wolfe Algorithms

We develop parallel and distributed Frank-Wolfe algorithms; the former o...

Please sign up or login with your details

Forgot password? Click here to reset