Physics-informed reinforcement learning via probabilistic co-adjustment functions

09/11/2023
by   Nat Wannawas, et al.
0

Reinforcement learning of real-world tasks is very data inefficient, and extensive simulation-based modelling has become the dominant approach for training systems. However, in human-robot interaction and many other real-world settings, there is no appropriate one-model-for-all due to differences in individual instances of the system (e.g. different people) or necessary oversimplifications in the simulation models. This requires two approaches: 1. either learning the individual system's dynamics approximately from data which requires data-intensive training or 2. using a complete digital twin of the instances, which may not be realisable in many cases. We introduce two approaches: co-kriging adjustments (CKA) and ridge regression adjustment (RRA) as novel ways to combine the advantages of both approaches. Our adjustment methods are based on an auto-regressive AR1 co-kriging model that we integrate with GP priors. This yield a data- and simulation-efficient way of using simplistic simulation models (e.g., simple two-link model) and rapidly adapting them to individual instances (e.g., biomechanics of individual people). Using CKA and RRA, we obtain more accurate uncertainty quantification of the entire system's dynamics than pure GP-based and AR1 methods. We demonstrate the efficiency of co-kriging adjustment with an interpretable reinforcement learning control example, learning to control a biomechanical human arm using only a two-link arm simulation model (offline part) and CKA derived from a small amount of interaction data (on-the-fly online). Our method unlocks an efficient and uncertainty-aware way to implement reinforcement learning methods in real world complex systems for which only imperfect simulation models exist.

READ FULL TEXT
research
06/27/2022

When to Trust Your Simulator: Dynamics-Aware Hybrid Offline-and-Online Reinforcement Learning

Learning effective reinforcement learning (RL) policies to solve real-wo...
research
08/15/2019

Sample-efficient Deep Reinforcement Learning with Imaginary Rollouts for Human-Robot Interaction

Deep reinforcement learning has proven to be a great success in allowing...
research
02/26/2020

Efficient reinforcement learning control for continuum robots based on Inexplicit Prior Knowledge

Compared to rigid robots that are often studied in reinforcement learnin...
research
05/15/2023

Gaussian Process Port-Hamiltonian Systems: Bayesian Learning with Physics Prior

Data-driven approaches achieve remarkable results for the modeling of co...
research
07/22/2023

On-Robot Bayesian Reinforcement Learning for POMDPs

Robot learning is often difficult due to the expense of gathering data. ...
research
05/20/2023

Bridging Active Exploration and Uncertainty-Aware Deployment Using Probabilistic Ensemble Neural Network Dynamics

In recent years, learning-based control in robotics has gained significa...
research
03/17/2023

Towards AI-controlled FES-restoration of movements: Learning cycling stimulation pattern with reinforcement learning

Functional electrical stimulation (FES) has been increasingly integrated...

Please sign up or login with your details

Forgot password? Click here to reset