A Review of Off-Policy Evaluation in Reinforcement Learning

12/13/2022
by   Masatoshi Uehara, et al.
0

Reinforcement learning (RL) is one of the most vibrant research frontiers in machine learning and has been recently applied to solve a number of challenging problems. In this paper, we primarily focus on off-policy evaluation (OPE), one of the most fundamental topics in RL. In recent years, a number of OPE methods have been developed in the statistics and computer science literature. We provide a discussion on the efficiency bound of OPE, some of the existing state-of-the-art OPE methods, their statistical properties and some other related research directions that are currently actively explored.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
04/12/2021

Survey on reinforcement learning for language processing

In recent years some researchers have explored the use of reinforcement ...
research
03/03/2022

On Practical Reinforcement Learning: Provable Robustness, Scalability, and Statistical Efficiency

This thesis rigorously studies fundamental reinforcement learning (RL) m...
research
02/26/2022

Statistically Efficient Advantage Learning for Offline Reinforcement Learning in Infinite Horizons

We consider reinforcement learning (RL) methods in offline domains witho...
research
10/26/2022

A Bibliometric Analysis and Review on Reinforcement Learning for Transportation Applications

Transportation is the backbone of the economy and urban development. Imp...
research
01/07/2020

Reinforcement Learning via Fenchel-Rockafellar Duality

We review basic concepts of convex duality, focusing on the very general...
research
07/22/2021

Accelerating Quadratic Optimization with Reinforcement Learning

First-order methods for quadratic optimization such as OSQP are widely u...
research
07/20/2020

Lagrangian Duality in Reinforcement Learning

Although duality is used extensively in certain fields, such as supervis...

Please sign up or login with your details

Forgot password? Click here to reset