Hierarchical Representations and Explicit Memory: Learning Effective Navigation Policies on 3D Scene Graphs using Graph Neural Networks

08/02/2021
by   Zachary Ravichandran, et al.
5

Representations are crucial for a robot to learn effective navigation policies. Recent work has shown that mid-level perceptual abstractions, such as depth estimates or 2D semantic segmentation, lead to more effective policies when provided as observations in place of raw sensor data (e.g., RGB images). However, such policies must still learn latent three-dimensional scene properties from mid-level abstractions. In contrast, high-level, hierarchical representations such as 3D scene graphs explicitly provide a scene's geometry, topology, and semantics, making them compelling representations for navigation. In this work, we present a reinforcement learning framework that leverages high-level hierarchical representations to learn navigation policies. Towards this goal, we propose a graph neural network architecture and show how to embed a 3D scene graph into an agent-centric feature space, which enables the robot to learn policies for low-level action in an end-to-end manner. For each node in the scene graph, our method uses features that capture occupancy and semantic content, while explicitly retaining memory of the robot trajectory. We demonstrate the effectiveness of our method against commonly used visuomotor policies in a challenging object search task. These experiments and supporting ablation studies show that our method leads to more effective object search behaviors, exhibits improved long-term memory, and successfully leverages hierarchical information to guide its navigation objectives.

READ FULL TEXT

page 4

page 6

page 16

page 17

page 18

page 19

page 20

research
12/14/2020

Hierarchical Planning for Long-Horizon Manipulation with Geometric and Symbolic Scene Graphs

We present a visually grounded hierarchical planning algorithm for long-...
research
08/17/2020

Learning to Actively Reduce Memory Requirements for Robot Control Tasks

Robots equipped with rich sensing modalities (e.g., RGB-D cameras) perfo...
research
11/21/2018

Early Fusion for Goal Directed Robotic Vision

Increasingly, perceptual systems are being codified as strict pipelines ...
research
11/15/2019

A Policy Editor for Semantic Sensor Networks

An important use of sensors and actuator networks is to comply with heal...
research
01/06/2023

ReVoLT: Relational Reasoning and Voronoi Local Graph Planning for Target-driven Navigation

Embodied AI is an inevitable trend that emphasizes the interaction betwe...
research
12/08/2022

Latent Graph Representations for Critical View of Safety Assessment

Assessing the critical view of safety in laparoscopic cholecystectomy re...
research
03/10/2019

Affordance Learning for End-to-End Visuomotor Robot Control

Training end-to-end deep robot policies requires a lot of domain-, task-...

Please sign up or login with your details

Forgot password? Click here to reset