Convergence of Variance-Reduced Stochastic Learning under Random Reshuffling

08/04/2017
by   Bicheng Ying, et al.
0

Several useful variance-reduced stochastic gradient algorithms, such as SVRG, SAGA, Finito, and SAG, have been proposed to minimize empirical risks with linear convergence properties to the exact minimizers. The existing convergence results assume uniform data sampling with replacement. However, it has been observed that random reshuffling can deliver superior performance. No formal proofs or guarantees of exact convergence exist for variance-reduced algorithms under random reshuffling. This paper resolves this open convergence issue and provides the first theoretical guarantee of linear convergence under random reshuffling for SAGA; the argument is also adaptable to other variance-reduced algorithms. Under random reshuffling, the paper further proposes a new amortized variance-reduced gradient (AVRG) algorithm with constant storage requirements compared to SAGA and with balanced gradient computations compared to SVRG. The balancing in computations are attained by amortizing the full gradient calculation across all iterations. AVRG is also shown analytically to converge linearly.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
11/08/2019

Variance Reduced Stochastic Proximal Algorithm for AUC Maximization

Stochastic Gradient Descent has been widely studied with classification ...
research
02/19/2021

A Variance Controlled Stochastic Method with Biased Estimation for Faster Non-convex Optimization

In this paper, we proposed a new technique, variance controlled stochast...
research
01/24/2019

Don't Jump Through Hoops and Remove Those Loops: SVRG and Katyusha are Better Without the Outer Loop

The stochastic variance-reduced gradient method (SVRG) and its accelerat...
research
11/20/2018

Variance Reduction in Stochastic Particle-Optimization Sampling

Stochastic particle-optimization sampling (SPOS) is a recently-developed...
research
03/21/2018

Stochastic Learning under Random Reshuffling

In empirical risk optimization, it has been observed that stochastic gra...
research
03/06/2020

Fast calculation of the variance of edge crossings in random linear arrangements

The interest in spatial networks where vertices are embedded in a one-di...
research
05/08/2022

Federated Random Reshuffling with Compression and Variance Reduction

Random Reshuffling (RR), which is a variant of Stochastic Gradient Desce...

Please sign up or login with your details

Forgot password? Click here to reset