Quadratic Gradient: Uniting Gradient Algorithm and Newton Method as One

09/03/2022
by   John Chiang, et al.
0

It might be inadequate for the line search technique for Newton's method to use only one floating point number. A column vector of the same size as the gradient might be better than a mere float number to accelerate each of the gradient elements with different rates. Moreover, a square matrix of the same order as the Hessian matrix might be helpful to correct the Hessian matrix. Chiang applied something between a column vector and a square matrix, namely a diagonal matrix, to accelerate the gradient and further proposed a faster gradient variant called quadratic gradient. In this paper, we present a new way to build a new version of the quadratic gradient. This new quadratic gradient doesn't satisfy the convergence conditions of the fixed Hessian Newton's method. However, experimental results show that it sometimes has a better performance than the original one in convergence rate. Also, Chiang speculates that there might be a relation between the Hessian matrix and the learning rate for the first-order gradient descent method. We prove that the floating number 1/ϵ + max{| λ_i | } can be a good learning rate of the gradient methods, where ϵ is a number to avoid division by zero and λ_i the eigenvalues of the Hessian matrix.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
12/20/2017

Block-diagonal Hessian-free Optimization for Training Neural Networks

Second-order methods for neural network optimization have several advant...
research
05/16/2023

The Average Rate of Convergence of the Exact Line Search Gradient Descent Method

It is very well-known that when the exact line search gradient descent m...
research
09/29/2020

Mathematical derivation for Vora-Value based filter design method: Gradient and Hessian

In this paper, we present the detailed mathematical derivation of the gr...
research
05/26/2019

Partial minimization of strict convex functions and tensor scaling

Assume that f is a strict convex function with a unique minimum in R^n. ...
research
07/21/2023

Newton Nonholonomic Source Seeking for Distance-Dependent Maps

The topics of source seeking and Newton-based extremum seeking have flou...
research
10/27/2019

Fast Evaluation and Approximation of the Gauss-Newton Hessian Matrix for the Multilayer Perceptron

We introduce a fast algorithm for entry-wise evaluation of the Gauss-New...

Please sign up or login with your details

Forgot password? Click here to reset