4DRVO-Net: Deep 4D Radar-Visual Odometry Using Multi-Modal and Multi-Scale Adaptive Fusion

08/12/2023
by   Guirong Zhuo, et al.
0

Four-dimensional (4D) radar–visual odometry (4DRVO) integrates complementary information from 4D radar and cameras, making it an attractive solution for achieving accurate and robust pose estimation. However, 4DRVO may exhibit significant tracking errors owing to three main factors: 1) sparsity of 4D radar point clouds; 2) inaccurate data association and insufficient feature interaction between the 4D radar and camera; and 3) disturbances caused by dynamic objects in the environment, affecting odometry estimation. In this paper, we present 4DRVO-Net, which is a method for 4D radar–visual odometry. This method leverages the feature pyramid, pose warping, and cost volume (PWC) network architecture to progressively estimate and refine poses. Specifically, we propose a multi-scale feature extraction network called Radar-PointNet++ that fully considers rich 4D radar point information, enabling fine-grained learning for sparse 4D radar point clouds. To effectively integrate the two modalities, we design an adaptive 4D radar–camera fusion module (A-RCFM) that automatically selects image features based on 4D radar point features, facilitating multi-scale cross-modal feature interaction and adaptive multi-modal feature fusion. In addition, we introduce a velocity-guided point-confidence estimation module to measure local motion patterns, reduce the influence of dynamic objects and outliers, and provide continuous updates during pose refinement. We demonstrate the excellent performance of our method and the effectiveness of each module design on both the VoD and in-house datasets. Our method outperforms all learning-based and geometry-based methods for most sequences in the VoD dataset. Furthermore, it has exhibited promising performance that closely approaches that of the 64-line LiDAR odometry results of A-LOAM without mapping optimization.

READ FULL TEXT

page 1

page 4

page 8

page 12

research
12/02/2020

PWCLO-Net: Deep LiDAR Odometry in 3D Point Clouds Using Hierarchical Embedding Mask Optimization

A novel 3D point cloud learning model for deep LiDAR odometry, named PWC...
research
11/03/2021

Efficient 3D Deep LiDAR Odometry

An efficient 3D point cloud learning architecture, named PWCLO-Net, for ...
research
08/25/2022

Bridging the View Disparity of Radar and Camera Features for Multi-modal Fusion 3D Object Detection

Environmental perception with multi-modal fusion of radar and camera is ...
research
05/11/2023

Multi-modal Multi-level Fusion for 3D Single Object Tracking

3D single object tracking plays a crucial role in computer vision. Mains...
research
12/23/2020

Warping of Radar Data into Camera Image for Cross-Modal Supervision in Automotive Applications

In this paper, we present a novel framework to project automotive radar ...
research
06/21/2022

What Goes Around: Leveraging a Constant-curvature Motion Constraint in Radar Odometry

This paper presents a method that leverages vehicle motion constraints t...
research
08/02/2020

Deep Visual Odometry with Adaptive Memory

We propose a novel deep visual odometry (VO) method that considers globa...

Please sign up or login with your details

Forgot password? Click here to reset