A Short Survey on Probabilistic Reinforcement Learning

01/21/2019
by   Reazul Hasan Russel, et al.
0

A reinforcement learning agent tries to maximize its cumulative payoff by interacting in an unknown environment. It is important for the agent to explore suboptimal actions as well as to pick actions with highest known rewards. Yet, in sensitive domains, collecting more data with exploration is not always possible, but it is important to find a policy with a certain performance guaranty. In this paper, we present a brief survey of methods available in the literature for balancing exploration-exploitation trade off and computing robust solutions from fixed samples in reinforcement learning.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
09/01/2021

A Survey of Exploration Methods in Reinforcement Learning

Exploration is an essential component of reinforcement learning algorith...
research
01/01/2020

Long-Term Visitation Value for Deep Exploration in Sparse Reward Reinforcement Learning

Reinforcement learning with sparse rewards is still an open challenge. C...
research
02/10/2021

Derivative-Free Reinforcement Learning: A Review

Reinforcement learning is about learning agent models that make the best...
research
01/23/2018

Curiosity-driven reinforcement learning with homeostatic regulation

We propose a curiosity reward based on information theory principles and...
research
06/13/2019

Curriculum Learning for Cumulative Return Maximization

Curriculum learning has been successfully used in reinforcement learning...
research
09/19/2013

Exploration and Exploitation in Visuomotor Prediction of Autonomous Agents

This paper discusses various techniques to let an agent learn how to pre...
research
05/30/2022

SEREN: Knowing When to Explore and When to Exploit

Efficient reinforcement learning (RL) involves a trade-off between "expl...

Please sign up or login with your details

Forgot password? Click here to reset