Variational Intrinsic Control

11/22/2016
by   Karol Gregor, et al.
0

In this paper we introduce a new unsupervised reinforcement learning method for discovering the set of intrinsic options available to an agent. This set is learned by maximizing the number of different states an agent can reliably reach, as measured by the mutual information between the set of options and option termination states. To this end, we instantiate two policy gradient based algorithms, one that creates an explicit embedding space of options and one that represents options implicitly. The algorithms also provide an explicit measure of empowerment in a given state that can be used by an empowerment maximizing agent. The algorithm scales well with function approximation and we demonstrate the applicability of the algorithm on a range of tasks.

READ FULL TEXT

page 5

page 6

page 9

page 10

research
10/06/2020

Diverse Exploration via InfoMax Options

In this paper, we study the problem of autonomously discovering temporal...
research
07/24/2019

Unsupervised Discovery of Decision States for Transfer in Reinforcement Learning

We present a hierarchical reinforcement learning (HRL) or options framew...
research
10/07/2020

Variational Intrinsic Control Revisited

In this paper, we revisit variational intrinsic control (VIC), an unsupe...
research
07/26/2018

Variational Option Discovery Algorithms

We explore methods for option discovery based on variational inference a...
research
01/01/2020

Options of Interest: Temporal Abstraction with Interest Functions

Temporal abstraction refers to the ability of an agent to use behaviours...
research
11/10/2017

Learning with Options that Terminate Off-Policy

A temporally abstract action, or an option, is specified by a policy and...
research
01/19/2020

Learning Options from Demonstration using Skill Segmentation

We present a method for learning options from segmented demonstration tr...

Please sign up or login with your details

Forgot password? Click here to reset