Group DETR v2: Strong Object Detector with Encoder-Decoder Pretraining

11/07/2022
by   Qiang Chen, et al.
0

We present a strong object detector with encoder-decoder pretraining and finetuning. Our method, called Group DETR v2, is built upon a vision transformer encoder ViT-Huge <cit.>, a DETR variant DINO <cit.>, and an efficient DETR training method Group DETR <cit.>. The training process consists of self-supervised pretraining and finetuning a ViT-Huge encoder on ImageNet-1K, pretraining the detector on Object365, and finally finetuning it on COCO. Group DETR v2 achieves 64.5 mAP on COCO test-dev, and establishes a new SoTA on the COCO leaderboard https://paperswithcode.com/sota/object-detection-on-coco

READ FULL TEXT

page 1

page 2

page 3

research
03/19/2021

Efficient Visual Pretraining with Contrastive Detection

Self-supervised pretraining has been shown to yield powerful representat...
research
08/08/2022

Label-Free Synthetic Pretraining of Object Detectors

We propose a new approach, Synthetic Optimized Layout with Instance Dete...
research
06/01/2022

MaskOCR: Text Recognition with Masked Encoder-Decoder Pretraining

In this paper, we present a model pretraining technique, named MaskOCR, ...
research
07/28/2023

SimDETR: Simplifying self-supervised pretraining for DETR

DETR-based object detectors have achieved remarkable performance but are...
research
05/23/2022

Decoder Denoising Pretraining for Semantic Segmentation

Semantic segmentation labels are expensive and time consuming to acquire...
research
11/04/2021

Bootstrap Your Object Detector via Mixed Training

We introduce MixTraining, a new training paradigm for object detection t...
research
10/29/2022

Pair DETR: Contrastive Learning Speeds Up DETR Training

The DETR object detection approach applies the transformer encoder and d...

Please sign up or login with your details

Forgot password? Click here to reset