Return-based Scaling: Yet Another Normalisation Trick for Deep RL

05/11/2021
by   Tom Schaul, et al.
0

Scaling issues are mundane yet irritating for practitioners of reinforcement learning. Error scales vary across domains, tasks, and stages of learning; sometimes by many orders of magnitude. This can be detrimental to learning speed and stability, create interference between learning tasks, and necessitate substantial tuning. We revisit this topic for agents based on temporal-difference learning, sketch out some desiderata and investigate scenarios where simple fixes fall short. The mechanism we propose requires neither tuning, clipping, nor adaptation. We validate its effectiveness and robustness on the suite of Atari games. Our scaling method turns out to be particularly helpful at mitigating interference, when training a shared neural network on multiple targets that differ in reward scale or discounting.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
09/06/2018

ANS: Adaptive Network Scaling for Deep Rectifier Reinforcement Learning Models

This work provides a thorough study on how reward scaling can affect per...
research
07/10/2023

Measuring and Mitigating Interference in Reinforcement Learning

Catastrophic interference is common in many network-based learning syste...
research
05/26/2019

Deep-Neural-Network based Fall-back Mechanism in Interference-Aware Receiver Design

In this letter, we consider designing a fall-back mechanism in an interf...
research
02/24/2016

Learning values across many orders of magnitude

Most learning algorithms are not invariant to the scale of the function ...
research
06/05/2022

Learning Dynamics and Generalization in Reinforcement Learning

Solving a reinforcement learning (RL) problem poses two competing challe...
research
04/29/2021

Adapting to Reward Progressivity via Spectral Reinforcement Learning

In this paper we consider reinforcement learning tasks with progressive ...

Please sign up or login with your details

Forgot password? Click here to reset