Point Transformer V2: Grouped Vector Attention and Partition-based Pooling

10/11/2022
by   Xiaoyang Wu, et al.
0

As a pioneering work exploring transformer architecture for 3D point cloud understanding, Point Transformer achieves impressive results on multiple highly competitive benchmarks. In this work, we analyze the limitations of the Point Transformer and propose our powerful and efficient Point Transformer V2 model with novel designs that overcome the limitations of previous work. In particular, we first propose group vector attention, which is more effective than the previous version of vector attention. Inheriting the advantages of both learnable weight encoding and multi-head attention, we present a highly effective implementation of grouped vector attention with a novel grouped weight encoding layer. We also strengthen the position information for attention by an additional position encoding multiplier. Furthermore, we design novel and lightweight partition-based pooling methods which enable better spatial alignment and more efficient sampling. Extensive experiments show that our model achieves better performance than its predecessor and achieves state-of-the-art on several challenging 3D point cloud understanding benchmarks, including 3D point cloud segmentation on ScanNet v2 and S3DIS and 3D point cloud classification on ModelNet40. Our code will be available at https://github.com/Gofinge/PointTransformerV2.

READ FULL TEXT
research
05/02/2023

PU-EdgeFormer: Edge Transformer for Dense Prediction in Point Cloud Upsampling

Despite the recent development of deep learning-based point cloud upsamp...
research
10/05/2022

Point Cloud Recognition with Position-to-Structure Attention Transformers

In this paper, we present Position-to-Structure Attention Transformers (...
research
03/23/2023

Position-Guided Point Cloud Panoptic Segmentation Transformer

DEtection TRansformer (DETR) started a trend that uses a group of learna...
research
06/28/2023

Tensorformer: Normalized Matrix Attention Transformer for High-quality Point Cloud Reconstruction

Surface reconstruction from raw point clouds has been studied for decade...
research
06/13/2023

Reinforcement Learning-Driven Linker Design via Fast Attention-based Point Cloud Alignment

Proteolysis-Targeting Chimeras (PROTACs) represent a novel class of smal...
research
02/13/2022

LighTN: Light-weight Transformer Network for Performance-overhead Tradeoff in Point Cloud Downsampling

Compared with traditional task-irrelevant downsampling methods, task-ori...
research
06/09/2022

PointNeXt: Revisiting PointNet++ with Improved Training and Scaling Strategies

PointNet++ is one of the most influential neural architectures for point...

Please sign up or login with your details

Forgot password? Click here to reset