Hierarchical Policy Learning is Sensitive to Goal Space Design

05/04/2019
by   Zach Dwiel, et al.
0

Hierarchy in reinforcement learning agents allows for control at multiple time scales yielding improved sample efficiency, the ability to deal with long time horizons and transferability of sub-policies to tasks outside the training distribution. It is often implemented as a master policy providing goals to a sub-policy. Ideally, we would like the goal-spaces to be learned, however, properties of optimal goal spaces still remain unknown and consequently there is no method yet to learn optimal goal spaces. Motivated by this, we systematically analyze how various modifications to the ground-truth goal-space affect learning in hierarchical models with the aim of identifying important properties of optimal goal spaces. Our results show that, while rotation of ground-truth goal spaces and noise had no effect, having additional unnecessary factors significantly impaired learning in hierarchical models.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/26/2018

Neural Modular Control for Embodied Question Answering

We present a modular approach for learning policies for navigation over ...
research
10/02/2018

Near-Optimal Representation Learning for Hierarchical Reinforcement Learning

We study the problem of representation learning in goal-conditioned hier...
research
05/20/2019

Reinforcement Learning without Ground-Truth State

To perform robot manipulation tasks, a low dimension state of the enviro...
research
09/20/2023

Hierarchical reinforcement learning with natural language subgoals

Hierarchical reinforcement learning has been a compelling approach for a...
research
05/13/2021

MapGo: Model-Assisted Policy Optimization for Goal-Oriented Tasks

In Goal-oriented Reinforcement learning, relabeling the raw goals in pas...
research
03/30/2020

Deep reinforcement learning for large-scale epidemic control

Epidemics of infectious diseases are an important threat to public healt...
research
11/17/2018

Autonomous Extraction of a Hierarchical Structure of Tasks in Reinforcement Learning, A Sequential Associate Rule Mining Approach

Reinforcement learning (RL) techniques, while often powerful, can suffer...

Please sign up or login with your details

Forgot password? Click here to reset