Investigating Recurrence and Eligibility Traces in Deep Q-Networks

04/18/2017
by   Jean Harb, et al.
0

Eligibility traces in reinforcement learning are used as a bias-variance trade-off and can often speed up training time by propagating knowledge back over time-steps in a single update. We investigate the use of eligibility traces in combination with recurrent networks in the Atari domain. We illustrate the benefits of both recurrent nets and eligibility traces in some Atari games, and highlight also the importance of the optimization used in the training.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/23/2018

Efficient Eligibility Traces for Deep Reinforcement Learning

Eligibility traces are an effective technique to accelerate reinforcemen...
research
04/03/2009

Eligibility Propagation to Speed up Time Hopping for Reinforcement Learning

A mechanism called Eligibility Propagation is proposed to speed up the T...
research
08/23/2020

Adaptive and Multiple Time-scale Eligibility Traces for Online Deep Reinforcement Learning

Deep reinforcement learning (DRL) is one of the promising approaches to ...
research
12/23/2021

Improving the Efficiency of Off-Policy Reinforcement Learning by Accounting for Past Decisions

Off-policy learning from multistep returns is crucial for sample-efficie...
research
01/26/2023

Trajectory-Aware Eligibility Traces for Off-Policy Reinforcement Learning

Off-policy learning from multistep returns is crucial for sample-efficie...
research
02/08/2019

Source Traces for Temporal Difference Learning

This paper motivates and develops source traces for temporal difference ...
research
06/04/2022

Explaining Hyperproperty Violations

Hyperproperties relate multiple computation traces to each other. Model ...

Please sign up or login with your details

Forgot password? Click here to reset