DPT: Deformable Patch-based Transformer for Visual Recognition

07/30/2021
by   Zhiyang Chen, et al.
0

Transformer has achieved great success in computer vision, while how to split patches in an image remains a problem. Existing methods usually use a fixed-size patch embedding which might destroy the semantics of objects. To address this problem, we propose a new Deformable Patch (DePatch) module which learns to adaptively split the images into patches with different positions and scales in a data-driven way rather than using predefined fixed patches. In this way, our method can well preserve the semantics in patches. The DePatch module can work as a plug-and-play module, which can easily be incorporated into different transformers to achieve an end-to-end training. We term this DePatch-embedded transformer as Deformable Patch-based Transformer (DPT) and conduct extensive evaluations of DPT on image classification and object detection. Results show DPT can achieve 81.9 classification, and 43.7 MSCOCO object detection. Code has been made available at: https://github.com/CASIA-IVA-Lab/DPT .

READ FULL TEXT

page 1

page 8

research
09/21/2022

PicT: A Slim Weakly Supervised Vision Transformer for Pavement Distress Classification

Automatic pavement distress classification facilitates improving the eff...
research
03/22/2022

Learning Patch-to-Cluster Attention in Vision Transformer

The Vision Transformer (ViT) model is built on the assumption of treatin...
research
04/11/2019

Detecting Repeating Objects using Patch Correlation Analysis

In this paper we describe a new method for detecting and counting a repe...
research
05/06/2023

DBAT: Dynamic Backward Attention Transformer for Material Segmentation with Cross-Resolution Patches

The objective of dense material segmentation is to identify the material...
research
04/10/2023

Use the Detection Transformer as a Data Augmenter

Detection Transformer (DETR) is a Transformer architecture based object ...
research
08/19/2022

Accelerating Vision Transformer Training via a Patch Sampling Schedule

We introduce the notion of a Patch Sampling Schedule (PSS), that varies ...
research
04/04/2022

BatchFormerV2: Exploring Sample Relationships for Dense Representation Learning

Attention mechanisms have been very popular in deep neural networks, whe...

Please sign up or login with your details

Forgot password? Click here to reset