PETR: Position Embedding Transformation for Multi-View 3D Object Detection

03/10/2022
by   Yingfei Liu, et al.
9

In this paper, we develop position embedding transformation (PETR) for multi-view 3D object detection. PETR encodes the position information of 3D coordinates into image features, producing the 3D position-aware features. Object query can perceive the 3D position-aware features and perform end-to-end object detection. PETR achieves state-of-the-art performance (50.4 44.1 It can serve as a simple yet strong baseline for future research.

READ FULL TEXT

page 7

page 13

page 14

research
03/17/2023

CAPE: Camera View Position Embedding for Multi-View 3D Object Detection

In this paper, we address the problem of detecting 3D objects from multi...
research
09/22/2021

MVM3Det: A Novel Method for Multi-view Monocular 3D Detection

Monocular 3D object detection encounters occlusion problems in many appl...
research
06/02/2022

PETRv2: A Unified Framework for 3D Perception from Multi-Camera Images

In this paper, we propose PETRv2, a unified framework for 3D perception ...
research
01/10/2023

FrustumFormer: Adaptive Instance-aware Resampling for Multi-view 3D Detection

The transformation of features from 2D perspective space to 3D space is ...
research
03/22/2016

Active Detection and Localization of Textureless Objects in Cluttered Environments

This paper introduces an active object detection and localization framew...
research
08/21/2023

QD-BEV : Quantization-aware View-guided Distillation for Multi-view 3D Object Detection

Multi-view 3D detection based on BEV (bird-eye-view) has recently achiev...
research
08/19/2022

Booster-SHOT: Boosting Stacked Homography Transformations for Multiview Pedestrian Detection with Attention

Improving multi-view aggregation is integral for multi-view pedestrian d...

Please sign up or login with your details

Forgot password? Click here to reset