Asynchronous Parallel Stochastic Gradient for Nonconvex Optimization

06/27/2015
by   Xiangru Lian, et al.
0

Asynchronous parallel implementations of stochastic gradient (SG) have been broadly used in solving deep neural network and received many successes in practice recently. However, existing theories cannot explain their convergence and speedup properties, mainly due to the nonconvexity of most deep learning formulations and the asynchronous parallel mechanism. To fill the gaps in theory and provide theoretical supports, this paper studies two asynchronous parallel implementations of SG: one is on the computer network and the other is on the shared memory system. We establish an ergodic convergence rate O(1/√(K)) for both algorithms and prove that the linear speedup is achievable if the number of workers is bounded by √(K) (K is the total number of iterations). Our results generalize and improve existing analysis for convex minimization.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/19/2018

A Model Parallel Proximal Stochastic Gradient Algorithm for Partially Asynchronous Systems

Large models are prevalent in modern machine learning scenarios, includi...
research
06/07/2018

Asynchronous Stochastic Quasi-Newton MCMC for Non-Convex Optimization

Recent studies have illustrated that stochastic gradient Markov Chain Mo...
research
01/12/2018

Asynchronous Stochastic Variational Inference

Stochastic variational inference (SVI) employs stochastic optimization t...
research
11/02/2020

Asynchronous Parallel Stochastic Quasi-Newton Methods

Although first-order stochastic algorithms, such as stochastic gradient ...
research
03/03/2021

Critical Parameters for Scalable Distributed Learning with Large Batches and Asynchronous Updates

It has been experimentally observed that the efficiency of distributed t...
research
11/08/2018

(Near) Optimal Parallelism Bound for Fully Asynchronous Coordinate Descent with Linear Speedup

When solving massive optimization problems in areas such as machine lear...
research
06/15/2016

ASAGA: Asynchronous Parallel SAGA

We describe ASAGA, an asynchronous parallel version of the incremental g...

Please sign up or login with your details

Forgot password? Click here to reset