Meta-Model-Based Meta-Policy Optimization

06/04/2020
by   Takuya Hiraoka, et al.
0

Model-based reinforcement learning (MBRL) has been applied to meta-learning settings and demonstrated its high sample efficiency. However, in previous MBRL for meta-learning settings, policies are optimized via rollouts that fully rely on a predictive model for an environment, and thus its performance in a real environment tends to degrade when the predictive model is inaccurate. In this paper, we prove that the performance degradation can be suppressed by using branched meta-rollouts. Based on this theoretical analysis, we propose meta-model-based meta-policy optimization (M3PO), in which the branched meta-rollouts are used for policy optimization. We demonstrate that M3PO outperforms existing meta reinforcement learning methods in continuous-control benchmarks.

READ FULL TEXT
research
11/21/2020

Double Meta-Learning for Data Efficient Policy Optimization in Non-Stationary Environments

We are interested in learning models of non-stationary environments, whi...
research
06/14/2020

Provably Efficient Model-based Policy Adaptation

The high sample complexity of reinforcement learning challenges its use ...
research
10/16/2018

ProMP: Proximal Meta-Policy Search

Credit assignment in Meta-reinforcement learning (Meta-RL) is still poor...
research
10/09/2020

Characterizing Policy Divergence for Personalized Meta-Reinforcement Learning

Despite ample motivation from costly exploration and limited trajectory ...
research
01/11/2021

Deep Interactive Bayesian Reinforcement Learning via Meta-Learning

Agents that interact with other agents often do not know a priori what t...
research
03/02/2021

Minimax Model Learning

We present a novel off-policy loss function for learning a transition mo...
research
02/01/2022

Tutorial on amortized optimization for learning to optimize over continuous domains

Optimization is a ubiquitous modeling tool that is often deployed in set...

Please sign up or login with your details

Forgot password? Click here to reset