A Stochastic Quasi-Newton Method with Nesterov's Accelerated Gradient

09/09/2019
by   S. Indrapriyadarsini, et al.
0

Incorporating second order curvature information in gradient based methods have shown to improve convergence drastically despite its computational intensity. In this paper, we propose a stochastic (online) quasi-Newton method with Nesterov's accelerated gradient in both its full and limited memory forms for solving large scale non-convex optimization problems in neural networks. The performance of the proposed algorithm is evaluated in Tensorflow on benchmark classification and regression problems. The results show improved performance compared to the classical second order oBFGS and oLBFGS methods and popular first order stochastic methods such as SGD and Adam. The performance with different momentum rates and batch sizes have also been illustrated.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/17/2019

A Stochastic Variance Reduced Nesterov's Accelerated Quasi-Newton Method

Recently algorithms incorporating second order curvature information hav...
research
10/21/2019

Implementation of a modified Nesterov's Accelerated quasi-Newton Method on Tensorflow

Recent studies incorporate Nesterov's accelerated gradient method for th...
research
09/29/2018

A fast quasi-Newton-type method for large-scale stochastic optimisation

During recent years there has been an increased interest in stochastic a...
research
05/01/2023

ISAAC Newton: Input-based Approximate Curvature for Newton's Method

We present ISAAC (Input-baSed ApproximAte Curvature), a novel method tha...
research
10/04/2016

A Generic Quasi-Newton Algorithm for Faster Gradient-Based Optimization

We propose a generic approach to accelerate gradient-based optimization ...
research
04/06/2020

Deep Neural Network Learning with Second-Order Optimizers – a Practical Study with a Stochastic Quasi-Gauss-Newton Method

Training in supervised deep learning is computationally demanding, and t...
research
12/14/2021

SC-Reg: Training Overparameterized Neural Networks under Self-Concordant Regularization

In this paper we propose the SC-Reg (self-concordant regularization) fra...

Please sign up or login with your details

Forgot password? Click here to reset