Accelerated Optimization Landscape of Linear-Quadratic Regulator

07/07/2023
by   Lechen Feng, et al.
0

Linear-quadratic regulator (LQR) is a landmark problem in the field of optimal control, which is the concern of this paper. Generally, LQR is classified into state-feedback LQR (SLQR) and output-feedback LQR (OLQR) based on whether the full state is obtained. It has been suggested in existing literature that both the SLQR and the OLQR could be viewed as constrained nonconvex matrix optimization problems in which the only variable to be optimized is the feedback gain matrix. In this paper, we introduce a first-order accelerated optimization framework of handling the LQR problem, and give its convergence analysis for the cases of SLQR and OLQR, respectively. Specifically, a Lipschiz Hessian property of LQR performance criterion is presented, which turns out to be a crucial property for the application of modern optimization techniques. For the SLQR problem, a continuous-time hybrid dynamic system is introduced, whose solution trajectory is shown to converge exponentially to the optimal feedback gain with Nesterov-optimal order 1-1/√(κ) (κ the condition number). Then, the symplectic Euler scheme is utilized to discretize the hybrid dynamic system, and a Nesterov-type method with a restarting rule is proposed that preserves the continuous-time convergence rate, i.e., the discretized algorithm admits the Nesterov-optimal convergence order. For the OLQR problem, a Hessian-free accelerated framework is proposed, which is a two-procedure method consisting of semiconvex function optimization and negative curvature exploitation. In a time 𝒪(ϵ^-7/4log(1/ϵ)), the method can find an ϵ-stationary point of the performance criterion; this entails that the method improves upon the 𝒪(ϵ^-2) complexity of vanilla gradient descent. Moreover, our method provides the second-order guarantee of stationary point.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
11/28/2017

Accelerated Gradient Descent Escapes Saddle Points Faster than Gradient Descent

Nesterov's accelerated gradient descent (AGD), an instance of the genera...
research
01/27/2022

Restarted Nonconvex Accelerated Gradient Descent: No More Polylogarithmic Factor in the O(ε^-7/4) Complexity

This paper studies the accelerated gradient descent for general nonconve...
research
12/11/2018

On the Curved Geometry of Accelerated Optimization

In this work we propose a differential geometric motivation for Nesterov...
research
12/26/2019

Convergence and sample complexity of gradient methods for the model-free linear quadratic regulator problem

Model-free reinforcement learning attempts to find an optimal control ac...
research
09/12/2022

On the Optimization Landscape of Dynamic Output Feedback: A Case Study for Linear Quadratic Regulator

The convergence of policy gradient algorithms in reinforcement learning ...
research
04/27/2022

Accelerated Continuous-Time Approximate Dynamic Programming via Data-Assisted Hybrid Control

We introduce a new closed-loop architecture for the online solution of a...
research
12/20/2018

Derivative-Free Methods for Policy Optimization: Guarantees for Linear Quadratic Systems

We study derivative-free methods for policy optimization over the class ...

Please sign up or login with your details

Forgot password? Click here to reset