Improving Levenberg-Marquardt Algorithm for Neural Networks

12/17/2022
by   Omead Pooladzandi, et al.
0

We explore the usage of the Levenberg-Marquardt (LM) algorithm for regression (non-linear least squares) and classification (generalized Gauss-Newton methods) tasks in neural networks. We compare the performance of the LM method with other popular first-order algorithms such as SGD and Adam, as well as other second-order algorithms such as L-BFGS , Hessian-Free and KFAC. We further speed up the LM method by using adaptive momentum, learning rate line search, and uphill step acceptance.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/28/2019

A Gram-Gauss-Newton Method Learning Overparameterized Deep Neural Networks for Regression Problems

First-order methods such as stochastic gradient descent (SGD) are curren...
research
06/20/2020

Training (Overparametrized) Neural Networks in Near-Linear Time

The slow convergence rate and pathological curvature issues of first-ord...
research
03/01/2023

AdaSAM: Boosting Sharpness-Aware Minimization with Adaptive Learning Rate and Momentum for Training Deep Neural Networks

Sharpness aware minimization (SAM) optimizer has been extensively explor...
research
12/14/2021

SC-Reg: Training Overparameterized Neural Networks under Self-Concordant Regularization

In this paper we propose the SC-Reg (self-concordant regularization) fra...
research
05/26/2022

Faster Optimization on Sparse Graphs via Neural Reparametrization

In mathematical optimization, second-order Newton's methods generally co...
research
05/23/2023

Layer-wise Adaptive Step-Sizes for Stochastic First-Order Methods for Deep Learning

We propose a new per-layer adaptive step-size procedure for stochastic f...
research
09/08/2021

Training Algorithm Matters for the Performance of Neural Network Potential

One hidden yet important issue for developing neural network potentials ...

Please sign up or login with your details

Forgot password? Click here to reset