A Regret Minimization Approach to Iterative Learning Control

02/26/2021
by   Naman Agarwal, et al.
11

We consider the setting of iterative learning control, or model-based policy learning in the presence of uncertain, time-varying dynamics. In this setting, we propose a new performance metric, planning regret, which replaces the standard stochastic uncertainty assumptions with worst case regret. Based on recent advances in non-stochastic control, we design a new iterative algorithm for minimizing planning regret that is more robust to model mismatch and uncertainty. We provide theoretical and empirical evidence that the proposed algorithm outperforms existing methods on several benchmarks.

READ FULL TEXT

page 1

page 2

page 3

page 4

06/22/2021

Regret-optimal Estimation and Control

We consider estimation and control in linear time-varying dynamical syst...
07/08/2020

Adaptive Regret for Control of Time-Varying Dynamics

We consider regret minimization for online control with time-varying lin...
06/06/2022

Learning to Control under Time-Varying Environment

This paper investigates the problem of regret minimization in linear tim...
07/13/2016

Safe Policy Improvement by Minimizing Robust Baseline Regret

An important problem in sequential decision-making under uncertainty is ...
04/01/2022

To Explore or Not to Explore: Regret-Based LTL Planning in Partially-Known Environments

In this paper, we investigate the optimal robot path planning problem fo...
08/18/2021

Scalable regret for learning to control network-coupled subsystems with unknown dynamics

We consider the problem of controlling an unknown linear quadratic Gauss...
06/10/2021

Differentiable Robust LQR Layers

This paper proposes a differentiable robust LQR layer for reinforcement ...