A Control-Theoretic Perspective on Optimal High-Order Optimization

12/16/2019
by   Tianyi Lin, et al.
0

In this paper, we provide a control-theoretic perspective on optimal tensor optimization algorithms for minimizing a convex function in a finite-dimensional Euclidean space. Given a function Φ: R^d →R that is convex and twice-continuously differentiable, we study an ordinary differential equation (ODE) that is governed by the gradient operator ∇Φ and a positive control parameter λ(t) that tends to infinity as t → +∞. The tuning of λ(·) is achieved via a closed-loop control law based on the algebraic equation[λ(t)]^p∇Φ(x(t))^p-1 = θ for a given θ > 0. We prove the existence and uniqueness of a local solution to this closed-loop ODE by the Banach fixed-point theorem. We then present a Lyapunov function that allows us to establish the existence and uniqueness of a global solution and analyze the convergence properties of trajectories. The rate of convergence is O(t^-(3p+1)/2) in terms of objective gap and O(t^-3p) in terms of squared gradient norm. We present two frameworks for implicit time discretization of the ODE, one of which generalizes the large-step A-HPE framework of <cit.>, and the other of which leads to a new p-th order tensor algorithm. A highlight of our analysis is that we show that all of the p-th order optimal tensor algorithms in this paper minimize the squared gradient norm at a rate of O(k^-3p).

READ FULL TEXT

page 1

page 2

page 3

page 4

research
11/15/2021

On Monotone Inclusions, Acceleration and Closed-Loop Control

We propose and analyze a new dynamical system with a closed-loop control...
research
06/09/2022

A Continuous-Time Perspective on Optimal Methods for Monotone Equation Problems

We study rescaled gradient dynamical systems in a Hilbert space ℋ, where...
research
04/13/2021

Approximating optimal feedback controllers of finite horizon control problems using hierarchical tensor formats

Controlling systems of ordinary differential equations (ODEs) is ubiquit...
research
10/04/2018

Convergence of the ADAM algorithm from a Dynamical System Viewpoint

Adam is a popular variant of the stochastic gradient descent for finding...
research
06/21/2012

Convergence of the Continuous Time Trajectories of Isotropic Evolution Strategies on Monotonic C^2-composite Functions

The Information-Geometric Optimization (IGO) has been introduced as a un...
research
12/31/2020

Optimizing Optimizers: Regret-optimal gradient descent algorithms

The need for fast and robust optimization algorithms are of critical imp...
research
02/12/2021

From perspective maps to epigraphical projections

The projection onto the epigraph or a level set of a closed proper conve...

Please sign up or login with your details

Forgot password? Click here to reset