Points to Patches: Enabling the Use of Self-Attention for 3D Shape Recognition

04/08/2022
by   Axel Berg, et al.
0

While the Transformer architecture has become ubiquitous in the machine learning field, its adaptation to 3D shape recognition is non-trivial. Due to its quadratic computational complexity, the self-attention operator quickly becomes inefficient as the set of input points grows larger. Furthermore, we find that the attention mechanism struggles to find useful connections between individual points on a global scale. In order to alleviate these problems, we propose a two-stage Point Transformer-in-Transformer (Point-TnT) approach which combines local and global attention mechanisms, enabling both individual points and patches of points to attend to each other effectively. Experiments on shape classification show that such an approach provides more useful features for downstream tasks than the baseline Transformer, while also being more computationally efficient. In addition, we also extend our method to feature matching for scene reconstruction, showing that it can be used in conjunction with existing scene reconstruction pipelines.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
03/29/2023

Self-positioning Point-based Transformer for Point Cloud Understanding

Transformers have shown superior performance on various computer vision ...
research
11/02/2020

Point Transformer

In this work, we present Point Transformer, a deep neural network that o...
research
09/05/2022

SEFormer: Structure Embedding Transformer for 3D Object Detection

Effectively preserving and encoding structure features from objects in i...
research
03/11/2023

Recursive Generalization Transformer for Image Super-Resolution

Transformer architectures have exhibited remarkable performance in image...
research
03/11/2022

Font Shape-to-Impression Translation

Different fonts have different impressions, such as elegant, scary, and ...
research
04/17/2020

Highway Transformer: Self-Gating Enhanced Self-Attentive Networks

Self-attention mechanisms have made striking state-of-the-art (SOTA) pro...
research
02/28/2023

Sampled Transformer for Point Sets

The sparse transformer can reduce the computational complexity of the se...

Please sign up or login with your details

Forgot password? Click here to reset