Perceive, Interact, Predict: Learning Dynamic and Static Clues for End-to-End Motion Prediction

12/05/2022
by   Bo Jiang, et al.
0

Motion prediction is highly relevant to the perception of dynamic objects and static map elements in the scenarios of autonomous driving. In this work, we propose PIP, the first end-to-end Transformer-based framework which jointly and interactively performs online mapping, object detection and motion prediction. PIP leverages map queries, agent queries and mode queries to encode the instance-wise information of map elements, agents and motion intentions, respectively. Based on the unified query representation, a differentiable multi-task interaction scheme is proposed to exploit the correlation between perception and prediction. Even without human-annotated HD map or agent's historical tracking trajectory as guidance information, PIP realizes end-to-end multi-agent motion prediction and achieves better performance than tracking-based and HD-map-based methods. PIP provides comprehensive high-level information of the driving scene (vectorized static map and dynamic objects with motion information), and contributes to the downstream planning and control. Code and models will be released for facilitating further research.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
03/21/2023

VAD: Vectorized Scene Representation for Efficient Autonomous Driving

Autonomous driving requires a comprehensive understanding of the surroun...
research
08/02/2022

ViP3D: End-to-end Visual Trajectory Prediction via 3D Agent Queries

Existing autonomous driving pipelines separate the perception module fro...
research
08/10/2023

MapTRv2: An End-to-End Framework for Online Vectorized HD Map Construction

High-definition (HD) map provides abundant and precise static environmen...
research
04/12/2022

Fully End-to-end Autonomous Driving with Semantic Depth Cloud Mapping and Multi-Agent

Focusing on the task of point-to-point navigation for an autonomous driv...
research
08/02/2023

Interpretable End-to-End Driving Model for Implicit Scene Understanding

Driving scene understanding is to obtain comprehensive scene information...
research
10/08/2020

BGM: Building a Dynamic Guidance Map without Visual Images for Trajectory Prediction

Visual images usually contain the informative context of the environment...
research
04/30/2022

HDGT: Heterogeneous Driving Graph Transformer for Multi-Agent Trajectory Prediction via Scene Encoding

One essential task for autonomous driving is to encode the information o...

Please sign up or login with your details

Forgot password? Click here to reset