PoseConvGRU: A Monocular Approach for Visual Ego-motion Estimation by Learning

06/19/2019
by   Guangyao Zhai, et al.
0

While many visual ego-motion algorithm variants have been proposed in the past decade, learning based ego-motion estimation methods have seen an increasing attention because of its desirable properties of robustness to image noise and camera calibration independence. In this work, we propose a data-driven approach of fully trainable visual ego-motion estimation for a monocular camera. We use an end-to-end learning approach in allowing the model to map directly from input image pairs to an estimate of ego-motion (parameterized as 6-DoF transformation matrices). We introduce a novel two-module Long-term Recurrent Convolutional Neural Networks called PoseConvGRU, with an explicit sequence pose estimation loss to achieve this. The feature-encoding module encodes the short-term motion feature in an image pair, while the memory-propagating module captures the long-term motion feature in the consecutive image pairs. The visual memory is implemented with convolutional gated recurrent units, which allows propagating information over time. At each time step, two consecutive RGB images are stacked together to form a 6 channels tensor for module-1 to learn how to extract motion information and estimate poses. The sequence of output maps is then passed through a stacked ConvGRU module to generate the relative transformation pose of each image pair. We also augment the training data by randomly skipping frames to simulate the velocity variation which results in a better performance in turning and high-velocity situations. We evaluate the performance of our proposed approach on the KITTI Visual Odometry benchmark. The experiments show a competitive performance of the proposed method to the geometric method and encourage further exploration of learning based methods for the purpose of estimating camera ego-motion even though geometrical methods demonstrate promising results.

READ FULL TEXT
research
05/05/2019

Learning by Inertia: Self-supervised Monocular Visual Odometry for Road Vehicles

In this paper, we present iDVO (inertia-embedded deep visual odometry), ...
research
12/19/2018

Deep Global-Relative Networks for End-to-End 6-DoF Visual Localization and Odometry

For the autonomous navigation of mobile robots, robust and fast visual l...
research
07/21/2020

Learning Monocular Visual Odometry via Self-Supervised Long-Term Modeling

Monocular visual odometry (VO) suffers severely from error accumulation ...
research
11/27/2018

MagicVO: End-to-End Monocular Visual Odometry through Deep Bi-directional Recurrent Convolutional Neural Network

This paper proposes a new framework to solve the problem of monocular vi...
research
11/25/2018

Guided Feature Selection for Deep Visual Odometry

We present a novel end-to-end visual odometry architecture with guided f...
research
07/29/2020

Deep Keypoint-Based Camera Pose Estimation with Geometric Constraints

Estimating relative camera poses from consecutive frames is a fundamenta...
research
08/03/2019

Learning Local Feature Descriptor with Motion Attribute for Vision-based Localization

In recent years, camera-based localization has been widely used for robo...

Please sign up or login with your details

Forgot password? Click here to reset