Variational Intrinsic Control Revisited

10/07/2020
by   Taehwan Kwon, et al.
0

In this paper, we revisit variational intrinsic control (VIC), an unsupervised reinforcement learning method for finding the largest set of intrinsic options available to an agent. In the original work by Gregor et al. (2016), two VIC algorithms were proposed: one that represents the options explicitly, and the other that does it implicitly. We show that the intrinsic reward used in the latter is subject to bias in stochastic environments, causing convergence to suboptimal solutions. To correct this behavior, we propose two methods respectively based on the transitional probability model and Gaussian Mixture Model. We substantiate our claims through rigorous mathematical derivations and experimental analyses.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
11/22/2016

Variational Intrinsic Control

In this paper we introduce a new unsupervised reinforcement learning met...
research
07/24/2019

Unsupervised Discovery of Decision States for Transfer in Reinforcement Learning

We present a hierarchical reinforcement learning (HRL) or options framew...
research
10/27/2018

Learning Abstract Options

Building systems that autonomously create temporal abstractions from dat...
research
01/16/2021

Hierarchical Reinforcement Learning By Discovering Intrinsic Options

We propose a hierarchical reinforcement learning method, HIDIO, that can...
research
10/11/2022

LECO: Learnable Episodic Count for Task-Specific Intrinsic Reward

Episodic count has been widely used to design a simple yet effective int...
research
06/16/2021

Revisiting the Weaknesses of Reinforcement Learning for Neural Machine Translation

Policy gradient algorithms have found wide adoption in NLP, but have rec...
research
06/09/2023

Computational Flash Photography through Intrinsics

Flash is an essential tool as it often serves as the sole controllable l...

Please sign up or login with your details

Forgot password? Click here to reset