Low-Precision Reinforcement Learning

02/26/2021
by   Johan Bjorck, et al.
51

Low-precision training has become a popular approach to reduce computation time, memory footprint, and energy consumption in supervised learning. In contrast, this promising approach has not enjoyed similarly widespread adoption within the reinforcement learning (RL) community, in part because RL agents can be notoriously hard to train – even in full precision. In this paper we consider continuous control with the state-of-the-art SAC agent and demonstrate that a naïve adaptation of low-precision methods from supervised learning fails. We propose a set of six modifications, all straightforward to implement, that leaves the underlying agent unchanged but improves its numerical stability dramatically. The resulting modified SAC agent has lower memory and compute requirements while matching full-precision rewards, thus demonstrating the feasibility of low-precision RL.

READ FULL TEXT

page 5

page 6

page 7

page 13

page 15

research
12/05/2019

Training Agents using Upside-Down Reinforcement Learning

Traditional Reinforcement Learning (RL) algorithms either predict reward...
research
08/10/2023

A Comparison of Classical and Deep Reinforcement Learning Methods for HVAC Control

Reinforcement learning (RL) is a promising approach for optimizing HVAC ...
research
10/13/2022

Personalized Federated Hypernetworks for Privacy Preservation in Multi-Task Reinforcement Learning

Multi-Agent Reinforcement Learning currently focuses on implementations ...
research
04/26/2019

Reinforcement Learning Based Orchestration for Elastic Services

Due to the highly variable execution context in which edge services run,...
research
03/14/2022

Constrained Precision Tuning

Precision tuning or customized precision number representations is emerg...
research
06/27/2012

Greedy Algorithms for Sparse Reinforcement Learning

Feature selection and regularization are becoming increasingly prominent...

Please sign up or login with your details

Forgot password? Click here to reset