Katyusha: The First Direct Acceleration of Stochastic Gradient Methods

03/18/2016
by   Zeyuan Allen-Zhu, et al.
0

Nesterov's momentum trick is famously known for accelerating gradient descent, and has been proven useful in building fast iterative algorithms. However, in the stochastic setting, counterexamples exist and prevent Nesterov's momentum from providing similar acceleration, even if the underlying problem is convex. We introduce Katyusha, a direct, primal-only stochastic gradient method to fix this issue. It has a provably accelerated convergence rate in convex (off-line) stochastic optimization. The main ingredient is Katyusha momentum, a novel "negative momentum" on top of Nesterov's momentum. It can be incorporated into a variance-reduction based algorithm and speed it up, both in terms of sequential and parallel performance. Since variance reduction has been successfully applied to a growing list of practical problems, our paper suggests that in each of such cases, one could potentially try to give Katyusha a hug.

READ FULL TEXT
research
06/28/2018

Direct Acceleration of SAGA using Sampled Negative Momentum

Variance reduction is a simple and effective technique that accelerates ...
research
06/19/2020

How Does Momentum Help Frank Wolfe?

We unveil the connections between Frank Wolfe (FW) type algorithms and t...
research
09/28/2021

An Accelerated Stochastic Gradient for Canonical Polyadic Decomposition

We consider the problem of structured canonical polyadic decomposition. ...
research
07/10/2017

Accelerated Stochastic Power Iteration

Principal component analysis (PCA) is one of the most powerful tools in ...
research
10/02/2020

Accelerating Convergence of Replica Exchange Stochastic Gradient MCMC via Variance Reduction

Replica exchange stochastic gradient Langevin dynamics (reSGLD) has show...
research
09/30/2022

Momentum Tracking: Momentum Acceleration for Decentralized Deep Learning on Heterogeneous Data

SGD with momentum acceleration is one of the key components for improvin...
research
02/10/2023

Achieving acceleration despite very noisy gradients

We present a novel momentum-based first order optimization method (AGNES...

Please sign up or login with your details

Forgot password? Click here to reset