How to Train PointGoal Navigation Agents on a (Sample and Compute) Budget

12/11/2020
by   Erik Wijmans, et al.
14

PointGoal navigation has seen significant recent interest and progress, spurred on by the Habitat platform and associated challenge. In this paper, we study PointGoal navigation under both a sample budget (75 million frames) and a compute budget (1 GPU for 1 day). We conduct an extensive set of experiments, cumulatively totaling over 50,000 GPU-hours, that let us identify and discuss a number of ostensibly minor but significant design choices – the advantage estimation procedure (a key component in training), visual encoder architecture, and a seemingly minor hyper-parameter change. Overall, these design choices to lead considerable and consistent improvements over the baselines present in Savva et al. Under a sample budget, performance for RGB-D agents improves 8 SPL on Gibson (14 Matterport3D (38 for RGB-D agents improves by 19 SPL on Gibson (32 SPL on Matterport3D (220 recommendations will make serve to make the community's experiments more efficient.

READ FULL TEXT

page 4

page 5

page 6

page 7

page 8

page 9

page 10

page 13

research
07/09/2020

Auxiliary Tasks Speed Up Learning PointGoal Navigation

PointGoal Navigation is an embodied task that requires agents to navigat...
research
07/24/2020

BabyAI 1.1

The BabyAI platform is designed to measure the sample efficiency of trai...
research
04/28/2021

Pushing it out of the Way: Interactive Visual Navigation

We have observed significant progress in visual navigation for embodied ...
research
06/08/2021

RobustNav: Towards Benchmarking Robustness in Embodied Navigation

As an attempt towards assessing the robustness of embodied navigation ag...
research
10/01/2021

Adwords with Unknown Budgets

Motivated by applications in automated budget optimization, we consider ...
research
05/30/2023

Bigger, Better, Faster: Human-level Atari with human-level efficiency

We introduce a value-based RL agent, which we call BBF, that achieves su...
research
03/02/2023

The Ladder in Chaos: A Simple and Effective Improvement to General DRL Algorithms by Policy Path Trimming and Boosting

Knowing the learning dynamics of policy is significant to unveiling the ...

Please sign up or login with your details

Forgot password? Click here to reset