Semantics-guided Transformer-based Sensor Fusion for Improved Waypoint Prediction

08/04/2023
by   Hwan-Soo Choi, et al.
0

Sensor fusion approaches for intelligent self-driving agents remain key to driving scene understanding given visual global contexts acquired from input sensors. Specifically, for the local waypoint prediction task, single-modality networks are still limited by strong dependency on the sensitivity of the input sensor, and thus recent works promote the use of multiple sensors in fusion in feature level. While it is well known that multiple data modalities promote mutual contextual exchange, deployment to practical driving scenarios requires global 3D scene understanding in real-time with minimal computations, thus placing greater significance on training strategies given a limited number of practically usable sensors. In this light, we exploit carefully selected auxiliary tasks that are highly correlated with the target task of interest (e.g., traffic light recognition and semantic segmentation) by fusing auxiliary task features and also using auxiliary heads for waypoint prediction based on imitation learning. Our multi-task feature fusion augments and improves the base network, TransFuser, by significant margins for safer and more complete road navigation in CARLA simulator as validated on the Town05 Benchmark through extensive experiments.

READ FULL TEXT

page 2

page 4

research
04/19/2021

Multi-Modal Fusion Transformer for End-to-End Autonomous Driving

How should representations from complementary sensors be integrated for ...
research
05/31/2022

TransFuser: Imitation with Transformer-Based Sensor Fusion for Autonomous Driving

How should we integrate representations from complementary sensors for a...
research
03/21/2023

Penalty-Based Imitation Learning With Cross Semantics Generation Sensor Fusion for Autonomous Driving

With the rapid development of Pattern Recognition and Computer Vision te...
research
06/30/2022

HRFuser: A Multi-resolution Sensor Fusion Architecture for 2D Object Detection

Besides standard cameras, autonomous vehicles typically include multiple...
research
08/17/2023

Sensor Fusion by Spatial Encoding for Autonomous Driving

Sensor fusion is critical to perception systems for task domains such as...
research
08/04/2023

Multi-interactive Feature Learning and a Full-time Multi-modality Benchmark for Image Fusion and Segmentation

Multi-modality image fusion and segmentation play a vital role in autono...
research
01/09/2021

Investigating the Effect of Sensor Modalities in Multi-Sensor Detection-Prediction Models

Detection of surrounding objects and their motion prediction are critica...

Please sign up or login with your details

Forgot password? Click here to reset