Tracking the gradients using the Hessian: A new look at variance reducing stochastic methods

10/20/2017
by   Robert M. Gower, et al.
0

Our goal is to improve variance reducing stochastic methods through better control variates. We first propose a modification of SVRG which uses the Hessian to track gradients over time, rather than to recondition, increasing the correlation of the control variates and leading to faster theoretical convergence close to the optimum. We then propose accurate and computationally efficient approximations to the Hessian, both using a diagonal and a low-rank matrix. Finally, we demonstrate the effectiveness of our method on a wide range of problems.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/08/2019

Reducing the variance in online optimization by transporting past gradients

Most stochastic optimization methods use gradients once before discardin...
research
10/08/2020

Dissecting Hessian: Understanding Common Structure of Hessian in Neural Networks

Hessian captures important properties of the deep neural network loss la...
research
06/16/2020

GO Hessian for Expectation-Based Objectives

An unbiased low-variance gradient estimator, termed GO gradient, was pro...
research
04/18/2023

Hessian and increasing-Hessian orderings of multivariate skew-elliptical random vectors

In this work, we establish some stochastic comparison results for multiv...
research
02/07/2020

Low Rank Saddle Free Newton: Algorithm and Analysis

Many tasks in engineering fields and machine learning involve minimizing...
research
06/27/2012

Estimating the Hessian by Back-propagating Curvature

In this work we develop Curvature Propagation (CP), a general technique ...
research
06/09/2023

Estimation of Ridge Using Nonlinear Transformation on Density Function

Ridges play a vital role in accurately approximating the underlying stru...

Please sign up or login with your details

Forgot password? Click here to reset