An Equivalence between Loss Functions and Non-Uniform Sampling in Experience Replay

07/12/2020
by   Scott Fujimoto, et al.
0

Prioritized Experience Replay (PER) is a deep reinforcement learning technique in which agents learn from transitions sampled with non-uniform probability proportionate to their temporal-difference error. We show that any loss function evaluated with non-uniformly sampled data can be transformed into another uniformly sampled loss function with the same expected gradient. Surprisingly, we find in some environments PER can be replaced entirely by this new loss function without impact to empirical performance. Furthermore, this relationship suggests a new branch of improvements to PER by correcting its uniformly sampled loss function equivalent. We demonstrate the effectiveness of our proposed modifications to PER and the equivalent loss function in several MuJoCo and Atari environments.

READ FULL TEXT
research
02/22/2021

Stratified Experience Replay: Correcting Multiplicity Bias in Off-Policy Reinforcement Learning

Deep Reinforcement Learning (RL) methods rely on experience replay to ap...
research
09/13/2023

Attention Loss Adjusted Prioritized Experience Replay

Prioritized Experience Replay (PER) is a technical means of deep reinfor...
research
04/05/2020

Dynamic Decision Boundary for One-class Classifiers applied to non-uniformly Sampled Data

A typical issue in Pattern Recognition is the non-uniformly sampled data...
research
04/23/2018

State Distribution-aware Sampling for Deep Q-learning

A critical and challenging problem in reinforcement learning is how to l...
research
02/05/2021

The Fourier Loss Function

This paper introduces a new loss function induced by the Fourier-based M...
research
08/08/2023

Scope Loss for Imbalanced Classification and RL Exploration

We demonstrate equivalence between the reinforcement learning problem an...

Please sign up or login with your details

Forgot password? Click here to reset