Contrastive Learning of Features between Images and LiDAR

06/24/2022
by   Peng Jiang, et al.
0

Image and Point Clouds provide different information for robots. Finding the correspondences between data from different sensors is crucial for various tasks such as localization, mapping, and navigation. Learning-based descriptors have been developed for single sensors; there is little work on cross-modal features. This work treats learning cross-modal features as a dense contrastive learning problem. We propose a Tuple-Circle loss function for cross-modality feature learning. Furthermore, to learn good features and not lose generality, we developed a variant of widely used PointNet++ architecture for point cloud and U-Net CNN architecture for images. Moreover, we conduct experiments on a real-world dataset to show the effectiveness of our loss function and network structure. We show that our models indeed learn information from both images as well as LiDAR by visualizing the features.

READ FULL TEXT

page 3

page 5

page 6

research
07/20/2022

Cross-Modal Contrastive Representation Learning for Audio-to-Image Generation

Multiple modalities for certain information provide a variety of perspec...
research
04/14/2022

CroCo: Cross-Modal Contrastive learning for localization of Earth Observation data

It is of interest to localize a ground-based LiDAR point cloud on remote...
research
12/06/2022

Attention-Enhanced Cross-modal Localization Between 360 Images and Point Clouds

Visual localization plays an important role for intelligent robots and a...
research
04/17/2023

(LC)^2: LiDAR-Camera Loop Constraints For Cross-Modal Place Recognition

Localization has been a challenging task for autonomous navigation. A lo...
research
05/28/2020

Self-supervised Modal and View Invariant Feature Learning

Most of the existing self-supervised feature learning methods for 3D dat...
research
03/02/2022

X-Trans2Cap: Cross-Modal Knowledge Transfer using Transformer for 3D Dense Captioning

3D dense captioning aims to describe individual objects by natural langu...
research
04/26/2018

Boosting LiDAR-based Semantic Labeling by Cross-Modal Training Data Generation

Mobile robots and autonomous vehicles rely on multi-modal sensor setups ...

Please sign up or login with your details

Forgot password? Click here to reset