A Regret Minimization Approach to Iterative Learning Control

02/26/2021
by   Naman Agarwal, et al.
11

We consider the setting of iterative learning control, or model-based policy learning in the presence of uncertain, time-varying dynamics. In this setting, we propose a new performance metric, planning regret, which replaces the standard stochastic uncertainty assumptions with worst case regret. Based on recent advances in non-stochastic control, we design a new iterative algorithm for minimizing planning regret that is more robust to model mismatch and uncertainty. We provide theoretical and empirical evidence that the proposed algorithm outperforms existing methods on several benchmarks.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/22/2021

Regret-optimal Estimation and Control

We consider estimation and control in linear time-varying dynamical syst...
research
07/08/2020

Adaptive Regret for Control of Time-Varying Dynamics

We consider regret minimization for online control with time-varying lin...
research
06/06/2022

Learning to Control under Time-Varying Environment

This paper investigates the problem of regret minimization in linear tim...
research
04/12/2023

Optimizing Sensor Allocation against Attackers with Uncertain Intentions: A Worst-Case Regret Minimization Approach

This paper is concerned with the optimal allocation of detection resourc...
research
07/13/2016

Safe Policy Improvement by Minimizing Robust Baseline Regret

An important problem in sequential decision-making under uncertainty is ...
research
04/01/2022

To Explore or Not to Explore: Regret-Based LTL Planning in Partially-Known Environments

In this paper, we investigate the optimal robot path planning problem fo...
research
06/10/2021

Differentiable Robust LQR Layers

This paper proposes a differentiable robust LQR layer for reinforcement ...

Please sign up or login with your details

Forgot password? Click here to reset