Fine-Grained Vehicle Perception via 3D Part-Guided Visual Data Augmentation

12/15/2020
by   Feixiang Lu, et al.
6

Holistically understanding an object and its 3D movable parts through visual perception models is essential for enabling an autonomous agent to interact with the world. For autonomous driving, the dynamics and states of vehicle parts such as doors, the trunk, and the bonnet can provide meaningful semantic information and interaction states, which are essential to ensure the safety of the self-driving vehicle. Existing visual perception models mainly focus on coarse parsing such as object bounding box detection or pose estimation and rarely tackle these situations. In this paper, we address this important problem for autonomous driving by solving two critical issues using visual data augmentation. First, to deal with data scarcity, we propose an effective training data generation process by fitting a 3D car model with dynamic parts to vehicles in real images and then reconstructing human-vehicle interaction scenarios. This allows us to directly edit the real images using the aligned 3D parts, yielding effective training data generation for learning robust deep neural networks (DNNs). Second, to benchmark the quality of 3D part understanding, we collect a large dataset in real world driving scenarios with vehicles in uncommon states (VUS), i.e. with the door or trunk opened, etc. Experiments demonstrate our trained network with visual data augmentation largely outperforms other baselines in terms of 2D detection and instance segmentation accuracy. Our network yields large improvements in discovering and understanding these uncommon cases. Moreover, we plan to release all of the source code, the dataset, and the trained model on GitHub.

READ FULL TEXT

page 1

page 3

page 4

page 5

page 7

page 8

page 11

page 12

research
07/16/2020

PerMO: Perceiving More at Once from a Single Image for Autonomous Driving

We present a novel approach to detect, segment, and reconstruct complete...
research
07/15/2022

DOLPHINS: Dataset for Collaborative Perception enabled Harmonious and Interconnected Self-driving

Vehicle-to-Everything (V2X) network has enabled collaborative perception...
research
03/11/2021

Robust 2D/3D Vehicle Parsing in CVIS

We present a novel approach to robustly detect and perceive vehicles in ...
research
10/15/2021

DG-Labeler and DGL-MOTS Dataset: Boost the Autonomous Driving Perception

Multi-object tracking and segmentation (MOTS) is a critical task for aut...
research
07/27/2021

Predicting Take-over Time for Autonomous Driving with Real-World Data: Robust Data Augmentation, Models, and Evaluation

Understanding occupant-vehicle interactions by modeling control transiti...
research
12/28/2022

Detection of Active Emergency Vehicles using Per-Frame CNNs and Output Smoothing

While inferring common actor states (such as position or velocity) is an...
research
03/18/2023

3D Data Augmentation for Driving Scenes on Camera

Driving scenes are extremely diverse and complicated that it is impossib...

Please sign up or login with your details

Forgot password? Click here to reset