Bigger, Better, Faster: Human-level Atari with human-level efficiency

05/30/2023
by   Max Schwarzer, et al.
0

We introduce a value-based RL agent, which we call BBF, that achieves super-human performance in the Atari 100K benchmark. BBF relies on scaling the neural networks used for value estimation, as well as a number of other design choices that enable this scaling in a sample-efficient manner. We conduct extensive analyses of these design choices and provide insights for future work. We end with a discussion about updating the goalposts for sample-efficient RL research on the ALE. We make our code and data publicly available at https://github.com/google-research/google-research/tree/master/bigger_better_faster.

READ FULL TEXT

page 6

page 7

page 8

research
03/31/2022

Scaling Up Models and Data with and

Recent neural network-based language models have benefited greatly from ...
research
09/07/2021

Learning Fast Sample Re-weighting Without Reward Data

Training sample re-weighting is an effective approach for tackling data ...
research
02/08/2021

Colorization Transformer

We present the Colorization Transformer, a novel approach for diverse hi...
research
01/31/2023

The Flan Collection: Designing Data and Methods for Effective Instruction Tuning

We study the design decisions of publicly available instruction tuning m...
research
10/18/2022

Helpful Neighbors: Leveraging Neighbors in Geographic Feature Pronunciation

If one sees the place name Houston Mercer Dog Run in New York, how does ...
research
12/11/2020

How to Train PointGoal Navigation Agents on a (Sample and Compute) Budget

PointGoal navigation has seen significant recent interest and progress, ...

Please sign up or login with your details

Forgot password? Click here to reset