Discrete State-Action Abstraction via the Successor Representation

06/07/2022
by   Amnon Attali, et al.
0

When reinforcement learning is applied with sparse rewards, agents must spend a prohibitively long time exploring the unknown environment without any learning signal. Abstraction is one approach that provides the agent with an intrinsic reward for transitioning in a latent space. Prior work focuses on dense continuous latent spaces, or requires the user to manually provide the representation. Our approach is the first for automatically learning a discrete abstraction of the underlying environment. Moreover, our method works on arbitrary input spaces, using an end-to-end trainable regularized successor representation model. For transitions between abstract states, we train a set of temporally extended actions in the form of options, i.e., an action abstraction. Our proposed algorithm, Discrete State-Action Abstraction (DSAA), iteratively swaps between training these options and using them to efficiently explore more of the environment to improve the state abstraction. As a result, our model is not only useful for transfer learning but also in the online learning setting. We empirically show that our agent is able to explore the environment and solve provided tasks more efficiently than baseline reinforcement learning algorithms. Our code is publicly available at <https://github.com/amnonattali/dsaa>.

READ FULL TEXT

page 5

page 10

page 15

research
08/08/2017

Investigating Reinforcement Learning Agents for Continuous State Space Environments

Given an environment with continuous state spaces and discrete actions, ...
research
03/09/2020

Learning discrete state abstractions with deep variational inference

Abstraction is crucial for effective sequential decision making in domai...
research
03/01/2022

A Theory of Abstraction in Reinforcement Learning

Reinforcement learning defines the problem facing agents that learn to m...
research
09/14/2022

A Simple Approach for State-Action Abstraction using a Learned MDP Homomorphism

Animals are able to rapidly infer from limited experience when sets of s...
research
10/05/2020

Randomized Value Functions via Posterior State-Abstraction Sampling

State abstraction has been an essential tool for dramatically improving ...
research
09/25/2022

Temporally Extended Successor Representations

We present a temporally extended variation of the successor representati...
research
04/06/2020

Uniform State Abstraction For Reinforcement Learning

Potential Based Reward Shaping combined with a potential function based ...

Please sign up or login with your details

Forgot password? Click here to reset