V2VNet: Vehicle-to-Vehicle Communication for Joint Perception and Prediction

by   Tsun-Hsuan Wang, et al.

In this paper, we explore the use of vehicle-to-vehicle (V2V) communication to improve the perception and motion forecasting performance of self-driving vehicles. By intelligently aggregating the information received from multiple nearby vehicles, we can observe the same scene from different viewpoints. This allows us to see through occlusions and detect actors at long range, where the observations are very sparse or non-existent. We also show that our approach of sending compressed deep feature map activations achieves high accuracy while satisfying communication bandwidth requirements.


Map Container: A Map-based Framework for Cooperative Perception

The idea of cooperative perception is to benefit from shared perception ...

Keypoints-Based Deep Feature Fusion for Cooperative Vehicle Detection of Autonomous Driving

Sharing collective perception messages (CPM) between vehicles is investi...

Vehicle to Vehicle (V2V) Communication Protocol: Components, Benefits, Challenges, Safety and Machine Learning Applications

Vehicle to vehicle communication is a new technology that enables vehicl...

DOLPHINS: Dataset for Collaborative Perception enabled Harmonious and Interconnected Self-driving

Vehicle-to-Everything (V2X) network has enabled collaborative perception...

Argoverse: 3D Tracking and Forecasting with Rich Maps

We present Argoverse – two datasets designed to support autonomous vehic...

Estimating Uncertainty of Autonomous Vehicle Systems with Generalized Polynomial Chaos

Modern autonomous vehicle systems use complex perception and control com...

MTBF Model for AVs – From Perception Errors to Vehicle-Level Failures

The development of Automated Vehicles (AVs) is progressing quickly and t...