A Conservation Law Method in Optimization

08/27/2017
by   Bin Shi, et al.
0

We propose some algorithms to find local minima in nonconvex optimization and to obtain global minima in some degree from the Newton Second Law without friction. With the key observation of the velocity observable and controllable in the motion, the algorithms simulate the Newton Second Law without friction based on symplectic Euler scheme. From the intuitive analysis of analytical solution, we give a theoretical analysis for the high-speed convergence in the algorithm proposed. Finally, we propose the experiments for strongly convex function, non-strongly convex function and nonconvex function in high-dimension.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
09/29/2022

On Quantum Speedups for Nonconvex Optimization via Quantum Tunneling Walks

Classical algorithms are often not effective for solving nonconvex optim...
research
04/05/2020

Regularized asymptotic descents for nonconvex optimization

In this paper we propose regularized asymptotic descent (RAD) methods fo...
research
04/05/2020

Regularized asymptotic descents for a class of nonconvex optimization problems

We propose and analyze regularized asymptotic descent (RAD) methods for ...
research
02/02/2017

Natasha: Faster Non-Convex Stochastic Optimization Via Strongly Non-Convex Parameter

Given a nonconvex function f(x) that is an average of n smooth functions...
research
01/10/2023

A Newton-CG based barrier-augmented Lagrangian method for general nonconvex conic optimization

In this paper we consider finding an approximate second-order stationary...
research
07/05/2021

The q-Levenberg-Marquardt method for unconstrained nonlinear optimization

A q-Levenberg-Marquardt method is an iterative procedure that blends a q...
research
07/26/2019

A simple Newton method for local nonsmooth optimization

Superlinear convergence has been an elusive goal for black-box nonsmooth...

Please sign up or login with your details

Forgot password? Click here to reset