Learning Deployable Navigation Policies at Kilometer Scale from a Single Traversal

07/11/2018
by   Jake Bruce, et al.
0

Model-free reinforcement learning has recently been shown to be effective at learning navigation policies from complex image input. However, these algorithms tend to require large amounts of interaction with the environment, which can be prohibitively costly to obtain on robots in the real world. We present an approach for efficiently learning goal-directed navigation policies on a mobile robot, from only a single coverage traversal of recorded data. The navigation agent learns an effective policy over a diverse action space in a large heterogeneous environment consisting of more than 2km of travel, through buildings and outdoor regions that collectively exhibit large variations in visual appearance, self-similarity, and connectivity. We compare pretrained visual encoders that enable precomputation of visual embeddings to achieve a throughput of tens of thousands of transitions per second at training time on a commodity desktop computer, allowing agents to learn from millions of trajectories of experience in a matter of hours. We propose multiple forms of computationally efficient stochastic augmentation to enable the learned policy to generalise beyond these precomputed embeddings, and demonstrate successful deployment of the learned policy on the real robot without fine tuning, despite environmental appearance differences at test time. The dataset and code required to reproduce these results and apply the technique to other datasets and robots is made publicly available at rl-navigation.github.io/deployable.

READ FULL TEXT

page 2

page 3

research
11/24/2020

Learning Navigation Skills for Legged Robots with Learned Robot Embeddings

Navigation policies are commonly learned on idealized cylinder agents in...
research
11/28/2017

One-Shot Reinforcement Learning for Robot Navigation with Interactive Replay

Recently, model-free reinforcement learning algorithms have been shown t...
research
10/14/2021

Augmenting Imitation Experience via Equivariant Representations

The robustness of visual navigation policies trained through imitation o...
research
10/10/2019

From Visual Place Recognition to Navigation: Learning Sample-Efficient Control Policies across Diverse Real World Environments

Visual navigation tasks in real world environments often require both se...
research
10/07/2022

GNM: A General Navigation Model to Drive Any Robot

Learning provides a powerful tool for vision-based navigation, but the c...
research
03/08/2022

Visual-Language Navigation Pretraining via Prompt-based Environmental Self-exploration

Vision-language navigation (VLN) is a challenging task due to its large ...
research
11/27/2018

Learning with Stochastic Guidance for Navigation

Due to the sparse rewards and high degree of environment variation, rein...

Please sign up or login with your details

Forgot password? Click here to reset