Adaptive Transformers in RL

04/08/2020
by   Shakti Kumar, et al.
0

Recent developments in Transformers have opened new interesting areas of research in partially observable reinforcement learning tasks. Results from late 2019 showed that Transformers are able to outperform LSTMs on both memory intense and reactive tasks. In this work we first partially replicate the results shown in Stabilizing Transformers in RL on both reactive and memory based environments. We then show performance improvement coupled with reduced computation when adding adaptive attention span to this Stable Transformer on a challenging DMLab30 environment. The code for all our experiments and models is available at https://github.com/jerrodparker20/adaptive-transformers-in-rl.

READ FULL TEXT
research
10/13/2019

Stabilizing Transformers for Reinforcement Learning

Owing to their ability to both effectively integrate information over lo...
research
01/09/2023

Transformers as Policies for Variable Action Environments

In this project we demonstrate the effectiveness of the transformer enco...
research
10/23/2020

Stabilizing Transformer-Based Action Sequence Generation For Q-Learning

Since the publication of the original Transformer architecture (Vaswani ...
research
03/03/2023

POPGym: Benchmarking Partially Observable Reinforcement Learning

Real world applications of Reinforcement Learning (RL) are often partial...
research
09/01/2022

Transformers are Sample Efficient World Models

Deep reinforcement learning agents are notoriously sample inefficient, w...
research
05/16/2023

Cooperation Is All You Need

Going beyond 'dendritic democracy', we introduce a 'democracy of local p...
research
06/11/2021

Going Beyond Linear Transformers with Recurrent Fast Weight Programmers

Transformers with linearised attention ("linear Transformers") have demo...

Please sign up or login with your details

Forgot password? Click here to reset