Saddle-free Hessian-free Optimization

05/30/2015
by   Martin Arjovsky, et al.
0

Nonconvex optimization problems such as the ones in training deep neural networks suffer from a phenomenon called saddle point proliferation. This means that there are a vast number of high error saddle points present in the loss function. Second order methods have been tremendously successful and widely adopted in the convex optimization community, while their usefulness in deep learning remains limited. This is due to two problems: computational complexity and the methods being driven towards the high error saddle points. We introduce a novel algorithm specially designed to solve these two issues, providing a crucial first step to take the widely known advantages of Newton's method to the nonconvex optimization community, especially in high dimensional settings.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
07/12/2022

A Newton-CG based barrier method for finding a second-order stationary point of nonconvex conic optimization with complexity guarantees

In this paper we consider finding an approximate second-order stationary...
research
11/28/2022

A survey of deep learning optimizers-first and second order methods

Deep Learning optimization involves minimizing a high-dimensional loss f...
research
12/02/2021

Newton methods based convolution neural networks using parallel processing

Training of convolutional neural networks is a high dimensional and a no...
research
09/26/2019

GradVis: Visualization and Second Order Analysis of Optimization Surfaces during the Training of Deep Neural Networks

Current training methods for deep neural networks boil down to very high...
research
07/31/2019

Multi-Point Bandit Algorithms for Nonstationary Online Nonconvex Optimization

Bandit algorithms have been predominantly analyzed in the convex setting...
research
07/09/2023

Large-scale global optimization of ultra-high dimensional non-convex landscapes based on generative neural networks

We present a non-convex optimization algorithm metaheuristic, based on t...
research
05/26/2017

Residual Expansion Algorithm: Fast and Effective Optimization for Nonconvex Least Squares Problems

We propose the residual expansion (RE) algorithm: a global (or near-glob...

Please sign up or login with your details

Forgot password? Click here to reset