Soft Actor-Critic for Discrete Action Settings

10/16/2019
by   Petros Christodoulou, et al.
0

Soft Actor-Critic is a state-of-the-art reinforcement learning algorithm for continuous action settings that is not applicable to discrete action settings. Many important settings involve discrete actions, however, and so here we derive an alternative version of the Soft Actor-Critic algorithm that is applicable to discrete action settings. We then show that, even without any hyperparameter tuning, it is competitive with the tuned model-free state-of-the-art on a selection of games from the Atari suite.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
09/21/2022

Revisiting Discrete Soft Actor-Critic

We study the adaption of soft actor-critic (SAC) from continuous action ...
research
03/11/2019

Sample-Efficient Model-Free Reinforcement Learning with Off-Policy Critics

Value-based reinforcement-learning algorithms are currently state-of-the...
research
09/17/2021

Soft Actor-Critic With Integer Actions

Reinforcement learning is well-studied under discrete actions. Integer a...
research
06/16/2021

Towards Automatic Actor-Critic Solutions to Continuous Control

Model-free off-policy actor-critic methods are an efficient solution to ...
research
12/06/2021

Target Entropy Annealing for Discrete Soft Actor-Critic

Soft Actor-Critic (SAC) is considered the state-of-the-art algorithm in ...
research
10/07/2019

Reinforcement Learning with Structured Hierarchical Grammar Representations of Actions

From a young age humans learn to use grammatical principles to hierarchi...
research
11/29/2019

Distributed Soft Actor-Critic with Multivariate Reward Representation and Knowledge Distillation

In this paper, we describe NeurIPS 2019 Learning to Move - Walk Around c...

Please sign up or login with your details

Forgot password? Click here to reset