Value Propagation Networks

05/28/2018
by   Nantas Nardelli, et al.
0

We present Value Propagation (VProp), a parameter-efficient differentiable planning module built on Value Iteration which can successfully be trained using reinforcement learning to solve unseen tasks, has the capability to generalize to larger map sizes, and can learn to navigate in dynamic environments. Furthermore, we show that the module enables learning to plan when the environment also includes stochastic elements, providing a cost-efficient learning system to build low-level size-invariant planners for a variety of interactive navigation problems. We evaluate on static and dynamic configurations of MazeBase grid-worlds, with randomly generated environments of several different sizes, and on a StarCraft navigation scenario, with more complex dynamics, and pixels as input.

READ FULL TEXT

page 5

page 8

research
02/09/2016

Value Iteration Networks

We introduce the value iteration network (VIN): a fully differentiable n...
research
08/08/2021

Towards real-world navigation with deep differentiable planners

We train embodied neural networks to plan and navigate unseen complex 3D...
research
06/08/2017

Generalized Value Iteration Networks: Life Beyond Lattices

In this paper, we introduce a generalized value iteration network (GVIN)...
research
06/07/2021

Hierarchical Robot Navigation in Novel Environments using Rough 2-D Maps

In robot navigation, generalizing quickly to unseen environments is esse...
research
05/11/2023

Value Iteration Networks with Gated Summarization Module

In this paper, we address the challenges faced by Value Iteration Networ...
research
10/16/2020

Robot Navigation in Constrained Pedestrian Environments using Reinforcement Learning

Navigating fluently around pedestrians is a necessary capability for mob...

Please sign up or login with your details

Forgot password? Click here to reset