A randomized primal distributed algorithm for partitioned and big-data non-convex optimization

03/24/2017
by   Ivano Notarnicola, et al.
0

In this paper we consider a distributed optimization scenario in which the aggregate objective function to minimize is partitioned, big-data and possibly non-convex. Specifically, we focus on a set-up in which the dimension of the decision variable depends on the network size as well as the number of local functions, but each local function handled by a node depends only on a (small) portion of the entire optimization variable. This problem set-up has been shown to appear in many interesting network application scenarios. As main paper contribution, we develop a simple, primal distributed algorithm to solve the optimization problem, based on a randomized descent approach, which works under asynchronous gossip communication. We prove that the proposed asynchronous algorithm is a proper, ad-hoc version of a coordinate descent method and thus converges to a stationary point. To show the effectiveness of the proposed algorithm, we also present numerical simulations on a non-convex quadratic program, which confirm the theoretical results.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
07/26/2021

Asynchronous Distributed Reinforcement Learning for LQR Control via Zeroth-Order Block Coordinate Descent

Recently introduced distributed zeroth-order optimization (ZOO) algorith...
research
10/24/2016

A Framework for Parallel and Distributed Training of Neural Networks

The aim of this paper is to develop a general framework for training neu...
research
10/13/2017

DSCOVR: Randomized Primal-Dual Block Coordinate Algorithms for Asynchronous Distributed Optimization

Machine learning with big data often involves large optimization models....
research
11/05/2019

Approximate Uncertain Program

Chance constrained program where one seeks to minimize an objective over...
research
12/10/2019

Byzantine Resilient Non-Convex SVRG with Distributed Batch Gradient Computations

In this work, we consider the distributed stochastic optimization proble...
research
10/08/2013

Distributed Coordinate Descent Method for Learning with Big Data

In this paper we develop and analyze Hydra: HYbriD cooRdinAte descent me...
research
07/09/2019

Randomized Constraints Consensus for Distributed Robust Mixed-Integer Programming

In this paper, we consider a network of processors aiming at cooperative...

Please sign up or login with your details

Forgot password? Click here to reset