A General Family of Robust Stochastic Operators for Reinforcement Learning

05/21/2018
by   Yingdong Lu, et al.
0

We consider a new family of operators for reinforcement learning with the goal of alleviating the negative effects and becoming more robust to approximation or estimation errors. Various theoretical results are established, which include showing on a sample path basis that our family of operators preserve optimality and increase the action gap. Our empirical results illustrate the strong benefits of our family of operators, significantly outperforming the classical Bellman operator and recently proposed operators.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
12/15/2015

Increasing the Action Gap: New Operators for Reinforcement Learning

This paper introduces new optimality-preserving operators on Q-functions...
research
11/13/2019

Approximation by Exponential Type Neural Network Operators

In the present article, we introduce and study the behaviour of the new ...
research
03/30/2022

Marginalized Operators for Off-policy Reinforcement Learning

In this work, we propose marginalized operators, a new class of off-poli...
research
04/22/2022

Aldaz-Kounchev-Render operators and their approximation properties

The approximation properties of the Aldaz-Kounchev-Render (AKR) operator...
research
05/02/2022

Incomplete Gamma Kernels: Generalizing Locally Optimal Projection Operators

We present incomplete gamma kernels, a generalization of Locally Optimal...
research
06/19/2021

Graph approximation and generalized Tikhonov regularization for signal deblurring

Given a compact linear operator , the (pseudo) inverse ^† is usually sub...
research
01/31/2018

A family of OWA operators based on Faulhaber's formulas

In this paper we develop a new family of Ordered Weighted Averaging (OWA...

Please sign up or login with your details

Forgot password? Click here to reset