Self-positioning Point-based Transformer for Point Cloud Understanding

03/29/2023
by   Jinyoung Park, et al.
0

Transformers have shown superior performance on various computer vision tasks with their capabilities to capture long-range dependencies. Despite the success, it is challenging to directly apply Transformers on point clouds due to their quadratic cost in the number of points. In this paper, we present a Self-Positioning point-based Transformer (SPoTr), which is designed to capture both local and global shape contexts with reduced complexity. Specifically, this architecture consists of local self-attention and self-positioning point-based global cross-attention. The self-positioning points, adaptively located based on the input shape, consider both spatial and semantic information with disentangled attention to improve expressive power. With the self-positioning points, we propose a novel global cross-attention mechanism for point clouds, which improves the scalability of global self-attention by allowing the attention module to compute attention weights with only a small set of self-positioning points. Experiments show the effectiveness of SPoTr on three point cloud tasks such as shape classification, part segmentation, and scene segmentation. In particular, our proposed model achieves an accuracy gain of 2.6 ScanObjectNN. We also provide qualitative analyses to demonstrate the interpretability of self-positioning points. The code of SPoTr is available at https://github.com/mlvlab/SPoTr.

READ FULL TEXT

page 4

page 8

research
06/02/2023

Collect-and-Distribute Transformer for 3D Point Cloud Analysis

Although remarkable advancements have been made recently in point cloud ...
research
07/31/2022

CloudAttention: Efficient Multi-Scale Attention Scheme For 3D Point Cloud Learning

Processing 3D data efficiently has always been a challenge. Spatial oper...
research
04/08/2022

Points to Patches: Enabling the Use of Self-Attention for 3D Shape Recognition

While the Transformer architecture has become ubiquitous in the machine ...
research
07/04/2023

DiT-3D: Exploring Plain Diffusion Transformers for 3D Shape Generation

Recent Diffusion Transformers (e.g., DiT) have demonstrated their powerf...
research
03/08/2023

Point Cloud Classification Using Content-based Transformer via Clustering in Feature Space

Recently, there have been some attempts of Transformer in 3D point cloud...
research
10/12/2018

PointGrow: Autoregressively Learned Point Cloud Generation with Self-Attention

A point cloud is an agile 3D representation, efficiently modeling an obj...
research
11/29/2020

Deeper or Wider Networks of Point Clouds with Self-attention?

Prevalence of deeper networks driven by self-attention is in stark contr...

Please sign up or login with your details

Forgot password? Click here to reset