Are We Ready for Vision-Centric Driving Streaming Perception? The ASAP Benchmark

12/17/2022
by   Xiaofeng Wang, et al.
0

In recent years, vision-centric perception has flourished in various autonomous driving tasks, including 3D detection, semantic map construction, motion forecasting, and depth estimation. Nevertheless, the latency of vision-centric approaches is too high for practical deployment (e.g., most camera-based 3D detectors have a runtime greater than 300ms). To bridge the gap between ideal research and real-world applications, it is necessary to quantify the trade-off between performance and efficiency. Traditionally, autonomous-driving perception benchmarks perform the offline evaluation, neglecting the inference time delay. To mitigate the problem, we propose the Autonomous-driving StreAming Perception (ASAP) benchmark, which is the first benchmark to evaluate the online performance of vision-centric perception in autonomous driving. On the basis of the 2Hz annotated nuScenes dataset, we first propose an annotation-extending pipeline to generate high-frame-rate labels for the 12Hz raw images. Referring to the practical deployment, the Streaming Perception Under constRained-computation (SPUR) evaluation protocol is further constructed, where the 12Hz inputs are utilized for streaming evaluation under the constraints of different computational resources. In the ASAP benchmark, comprehensive experiment results reveal that the model rank alters under different constraints, suggesting that the model latency and computation budget should be considered as design choices to optimize the practical deployment. To facilitate further research, we establish baselines for camera-based streaming 3D detection, which consistently enhance the streaming performance across various hardware. ASAP project page: https://github.com/JeffWang987/ASAP.

READ FULL TEXT

page 3

page 4

page 13

research
07/30/2021

Real-time Streaming Perception System for Autonomous Driving

Nowadays, plenty of deep learning technologies are being applied to all ...
research
07/14/2023

Linking vision and motion for self-supervised object-centric perception

Object-centric representations enable autonomous driving algorithms to r...
research
04/02/2023

One Training for Multiple Deployments: Polar-based Adaptive BEV Perception for Autonomous Driving

Current on-board chips usually have different computing power, which mea...
research
08/16/2022

Context-Aware Streaming Perception in Dynamic Environments

Efficient vision works maximize accuracy under a latency budget. These w...
research
09/13/2023

MTD: Multi-Timestep Detector for Delayed Streaming Perception

Autonomous driving systems require real-time environmental perception to...
research
09/13/2018

On Offline Evaluation of Vision-based Driving Models

Autonomous driving models should ideally be evaluated by deploying them ...
research
05/21/2020

Towards Streaming Image Understanding

Embodied perception refers to the ability of an autonomous agent to perc...

Please sign up or login with your details

Forgot password? Click here to reset