Modality-invariant Visual Odometry for Embodied Vision

04/29/2023
by   Marius Memmel, et al.
0

Effectively localizing an agent in a realistic, noisy setting is crucial for many embodied vision tasks. Visual Odometry (VO) is a practical substitute for unreliable GPS and compass sensors, especially in indoor environments. While SLAM-based methods show a solid performance without large data requirements, they are less flexible and robust w.r.t. to noise and changes in the sensor suite compared to learning-based approaches. Recent deep VO models, however, limit themselves to a fixed set of input modalities, e.g., RGB and depth, while training on millions of samples. When sensors fail, sensor suites change, or modalities are intentionally looped out due to available resources, e.g., power consumption, the models fail catastrophically. Furthermore, training these models from scratch is even more expensive without simulator access or suitable existing models that can be fine-tuned. While such scenarios get mostly ignored in simulation, they commonly hinder a model's reusability in real-world applications. We propose a Transformer-based modality-invariant VO approach that can deal with diverse or changing sensor suites of navigation agents. Our model outperforms previous methods while training on only a fraction of the data. We hope this method opens the door to a broader range of real-world applications that can benefit from flexible and learned VO models.

READ FULL TEXT

page 3

page 5

page 8

research
08/26/2021

The Surprising Effectiveness of Visual Odometry Techniques for Embodied PointGoal Navigation

It is fundamental for personal robots to reliably navigate to a specifie...
research
03/02/2020

MVP: Unified Motion and Visual Self-Supervised Learning for Large-Scale Robotic Navigation

Autonomous navigation emerges from both motion and local visual percepti...
research
09/07/2020

Integrating Egocentric Localization for More Realistic Point-Goal Navigation Agents

Recent work has presented embodied agents that can navigate to point-goa...
research
08/21/2020

Learning Camera-Aware Noise Models

Modeling imaging sensor noise is a fundamental problem for image process...
research
08/15/2018

A Comparative Study of Bug Algorithms for Robot Navigation

This paper presents a literature survey and a comparative study of Bug A...
research
01/30/2019

Benchmarking Classic and Learned Navigation in Complex 3D Environments

Navigation research is attracting renewed interest with the advent of le...
research
02/07/2020

Input Dropout for Spatially Aligned Modalities

Computer vision datasets containing multiple modalities such as color, d...

Please sign up or login with your details

Forgot password? Click here to reset