Globally Convergent Policy Search over Dynamic Filters for Output Estimation

02/23/2022
by   Jack Umenberger, et al.
0

We introduce the first direct policy search algorithm which provably converges to the globally optimal dynamic filter for the classical problem of predicting the outputs of a linear dynamical system, given noisy, partial observations. Despite the ubiquity of partial observability in practice, theoretical guarantees for direct policy search algorithms, one of the backbones of modern reinforcement learning, have proven difficult to achieve. This is primarily due to the degeneracies which arise when optimizing over filters that maintain internal state. In this paper, we provide a new perspective on this challenging problem based on the notion of informativity, which intuitively requires that all components of a filter's internal state are representative of the true state of the underlying dynamical system. We show that informativity overcomes the aforementioned degeneracy. Specifically, we propose a regularizer which explicitly enforces informativity, and establish that gradient descent on this regularized objective - combined with a “reconditioning step” - converges to the globally optimal cost a 𝒪(1/T) rate. Our analysis relies on several new results which may be of independent interest, including a new framework for analyzing non-convex gradient descent via convex reformulation, and novel bounds on the solution to linear Lyapunov equations in terms of (our quantitative measure of) informativity.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
09/16/2016

Gradient Descent Learns Linear Dynamical Systems

We prove that gradient descent efficiently converges to the global optim...
research
06/07/2022

Preconditioned Gradient Descent for Overparameterized Nonconvex Burer–Monteiro Factorization with Global Optimality Certification

We consider using gradient descent to minimize the nonconvex function f(...
research
05/26/2023

Fast and Minimax Optimal Estimation of Low-Rank Matrices via Non-Convex Gradient Descent

We study the problem of estimating a low-rank matrix from noisy measurem...
research
03/15/2023

Policy Gradient Converges to the Globally Optimal Policy for Nearly Linear-Quadratic Regulators

Nonlinear control systems with partial information to the decision maker...
research
12/26/2018

Dynamic Online Gradient Descent with Improved Query Complexity: A Theoretical Revisit

We provide a new theoretical analysis framework to investigate online gr...
research
12/08/2021

Reverse image filtering using total derivative approximation and accelerated gradient descent

In this paper, we address a new problem of reversing the effect of an im...
research
04/11/2017

Non-Linear Least-Squares Optimization of Rational Filters for the Solution of Interior Eigenvalue Problems

Rational filter functions can be used to improve convergence of contour-...

Please sign up or login with your details

Forgot password? Click here to reset