Image-based Navigation in Real-World Environments via Multiple Mid-level Representations: Fusion Models, Benchmark and Efficient Evaluation

02/02/2022
by   Marco Rosano, et al.
0

Navigating complex indoor environments requires a deep understanding of the space the robotic agent is acting into to correctly inform the navigation process of the agent towards the goal location. In recent learning-based navigation approaches, the scene understanding and navigation abilities of the agent are achieved simultaneously by collecting the required experience in simulation. Unfortunately, even if simulators represent an efficient tool to train navigation policies, the resulting models often fail when transferred into the real world. One possible solution is to provide the navigation model with mid-level visual representations containing important domain-invariant properties of the scene. But, what are the best representations that facilitate the transfer of a model to the real-world? How can they be combined? In this work we address these issues by proposing a benchmark of Deep Learning architectures to combine a range of mid-level visual representations, to perform a PointGoal navigation task following a Reinforcement Learning setup. All the proposed navigation models have been trained with the Habitat simulator on a synthetic office environment and have been tested on the same real-world environment using a real robotic platform. To efficiently assess their performance in a real context, a validation tool has been proposed to generate realistic navigation episodes inside the simulator. Our experiments showed that navigation models can benefit from the multi-modal input and that our validation tool can provide good estimation of the expected navigation performance in the real world, while saving time and resources. The acquired synthetic and real 3D models of the environment, together with the code of our validation tool built on top of Habitat, are publicly available at the following link: https://iplab.dmi.unict.it/EmbodiedVN/

READ FULL TEXT

page 3

page 6

page 8

page 9

page 13

page 14

page 17

page 18

research
10/26/2020

On Embodied Visual Navigation in Real Environments Through Habitat

Visual navigation models based on deep learning can learn effective poli...
research
05/12/2021

Out of the Box: Embodied Navigation in the Real World

The research field of Embodied AI has witnessed substantial progress in ...
research
12/11/2017

MINOS: Multimodal Indoor Simulator for Navigation in Complex Environments

We present MINOS, a simulator designed to support the development of mul...
research
01/08/2022

RARA: Zero-shot Sim2Real Visual Navigation with Following Foreground Cues

The gap between simulation and the real-world restrains many machine lea...
research
03/12/2021

Augmented Environment Representations with Complete Object Models

While 2D occupancy maps commonly used in mobile robotics enable safe nav...
research
08/24/2019

Situational Fusion of Visual Representation for Visual Navigation

A complex visual navigation task puts an agent in different situations w...
research
08/21/2023

Explore and Tell: Embodied Visual Captioning in 3D Environments

While current visual captioning models have achieved impressive performa...

Please sign up or login with your details

Forgot password? Click here to reset