SkipcrossNets: Adaptive Skip-cross Fusion for Road Detection

08/24/2023
by   Xinyu Zhang, et al.
0

Multi-modal fusion is increasingly being used for autonomous driving tasks, as images from different modalities provide unique information for feature extraction. However, the existing two-stream networks are only fused at a specific network layer, which requires a lot of manual attempts to set up. As the CNN goes deeper, the two modal features become more and more advanced and abstract, and the fusion occurs at the feature level with a large gap, which can easily hurt the performance. In this study, we propose a novel fusion architecture called skip-cross networks (SkipcrossNets), which combines adaptively LiDAR point clouds and camera images without being bound to a certain fusion epoch. Specifically, skip-cross connects each layer to each layer in a feed-forward manner, and for each layer, the feature maps of all previous layers are used as input and its own feature maps are used as input to all subsequent layers for the other modality, enhancing feature propagation and multi-modal features fusion. This strategy facilitates selection of the most similar feature layers from two data pipelines, providing a complementary effect for sparse point cloud features during fusion processes. The network is also divided into several blocks to reduce the complexity of feature fusion and the number of model parameters. The advantages of skip-cross fusion were demonstrated through application to the KITTI and A2D2 datasets, achieving a MaxF score of 96.85 parameters required only 2.33 MB of memory at a speed of 68.24 FPS, which could be viable for mobile terminals and embedded devices.

READ FULL TEXT

page 4

page 7

page 18

research
03/13/2023

A Generalized Multi-Modal Fusion Detection Framework

LiDAR point clouds have become the most common data source in autonomous...
research
03/15/2023

MSeg3D: Multi-modal 3D Semantic Segmentation for Autonomous Driving

LiDAR and camera are two modalities available for 3D semantic segmentati...
research
02/22/2022

Enabling Efficient Deep Convolutional Neural Network-based Sensor Fusion for Autonomous Driving

Autonomous driving demands accurate perception and safe decision-making....
research
07/17/2018

Robust Deep Multi-modal Learning Based on Gated Information Fusion Network

The goal of multi-modal learning is to use complimentary information on ...
research
09/21/2023

FGFusion: Fine-Grained Lidar-Camera Fusion for 3D Object Detection

Lidars and cameras are critical sensors that provide complementary infor...
research
07/16/2020

Memory Based Attentive Fusion

The use of multi-modal data for deep machine learning has shown promise ...
research
03/07/2020

Weight mechanism adding a constant in concatenation of series connect

It is a consensus that feature maps in the shallow layer are more relate...

Please sign up or login with your details

Forgot password? Click here to reset