Communication-Efficient Asynchronous Stochastic Frank-Wolfe over Nuclear-norm Balls

10/17/2019
by   Jiacheng Zhuo, et al.
17

Large-scale machine learning training suffers from two prior challenges, specifically for nuclear-norm constrained problems with distributed systems: the synchronization slowdown due to the straggling workers, and high communication costs. In this work, we propose an asynchronous Stochastic Frank Wolfe (SFW-asyn) method, which, for the first time, solves the two problems simultaneously, while successfully maintaining the same convergence rate as the vanilla SFW. We implement our algorithm in python (with MPI) to run on Amazon EC2, and demonstrate that SFW-asyn yields speed-ups almost linear to the number of machines compared to the vanilla SFW.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/21/2019

Sparsification as a Remedy for Staleness in Distributed Asynchronous SGD

Large scale machine learning is increasingly relying on distributed opti...
research
03/02/2020

BASGD: Buffered Asynchronous SGD for Byzantine Learning

Distributed learning has become a hot research topic, due to its wide ap...
research
02/01/2020

Finite-Time Analysis of Asynchronous Stochastic Approximation and Q-Learning

We consider a general asynchronous Stochastic Approximation (SA) scheme ...
research
01/12/2018

Asynchronous Stochastic Variational Inference

Stochastic variational inference (SVI) employs stochastic optimization t...
research
03/12/2019

Communication-efficient distributed SGD with Sketching

Large-scale distributed training of neural networks is often limited by ...
research
04/13/2017

Projection Free Rank-Drop Steps

The Frank-Wolfe (FW) algorithm has been widely used in solving nuclear n...
research
05/19/2018

Tell Me Something New: a new framework for asynchronous parallel learning

We present a novel approach for parallel computation in the context of m...

Please sign up or login with your details

Forgot password? Click here to reset