Exploration via Flow-Based Intrinsic Rewards

05/24/2019
by   Hsuan-Kung Yang, et al.
0

Exploration bonuses derived from the novelty of observations in an environment have become a popular approach to motivate exploration for reinforcement learning (RL) agents in the past few years. Recent methods such as curiosity-driven exploration usually estimate the novelty of new observations by the prediction errors of their system dynamics models. In this paper, we introduce the concept of optical flow estimation from the field of computer vision to the RL domain and utilize the errors from optical flow estimation to evaluate the novelty of new observations. We introduce a flow-based intrinsic curiosity module (FICM) capable of learning the motion features and understanding the observations in a more comprehensive and efficient fashion. We evaluate our method and compare it with a number of baselines on several benchmark environments, including Atari games, Super Mario Bros., and ViZDoom. Our results show that the proposed method is superior to the baselines in certain environments, especially for those featuring sophisticated moving patterns or with high-dimensional observation spaces. We further analyze the hyper-parameters used in the training phase and discuss our insights into them.

READ FULL TEXT

page 2

page 3

page 6

research
01/24/2019

Never Forget: Balancing Exploration and Exploitation via Learning Optical Flow

Exploration bonus derived from the novelty of the states in an environme...
research
04/21/2023

DEIR: Efficient and Robust Exploration through Discriminative-Model-Based Episodic Intrinsic Rewards

Exploration is a fundamental aspect of reinforcement learning (RL), and ...
research
09/19/2022

Rewarding Episodic Visitation Discrepancy for Exploration in Reinforcement Learning

Exploration is critical for deep reinforcement learning in complex envir...
research
08/09/2023

Intrinsic Motivation via Surprise Memory

We present a new computing model for intrinsic rewards in reinforcement ...
research
11/18/2022

Exploring through Random Curiosity with General Value Functions

Efficient exploration in reinforcement learning is a challenging problem...
research
03/22/2021

Transforming Exploratory Creativity with DeLeNoX

We introduce DeLeNoX (Deep Learning Novelty Explorer), a system that aut...
research
05/15/2023

MIMEx: Intrinsic Rewards from Masked Input Modeling

Exploring in environments with high-dimensional observations is hard. On...

Please sign up or login with your details

Forgot password? Click here to reset