Masked Autoencoders in 3D Point Cloud Representation Learning

07/04/2022
by   Jincen Jiang, et al.
0

Transformer-based Self-supervised Representation Learning methods learn generic features from unlabeled datasets for providing useful network initialization parameters for downstream tasks. Recently, self-supervised learning based upon masking local surface patches for 3D point cloud data has been under-explored. In this paper, we propose masked Autoencoders in 3D point cloud representation learning (abbreviated as MAE3D), a novel autoencoding paradigm for self-supervised learning. We first split the input point cloud into patches and mask a portion of them, then use our Patch Embedding Module to extract the features of unmasked patches. Secondly, we employ patch-wise MAE3D Transformers to learn both local features of point cloud patches and high-level contextual relationships between patches and complete the latent representations of masked patches. We use our Point Cloud Reconstruction Module with multi-task loss to complete the incomplete point cloud as a result. We conduct self-supervised pre-training on ShapeNet55 with the point cloud completion pre-text task and fine-tune the pre-trained model on ModelNet40 and ScanObjectNN (PB_T50_RS, the hardest variant). Comprehensive experiments demonstrate that the local features extracted by our MAE3D from point cloud patches are beneficial for downstream classification tasks, soundly outperforming state-of-the-art methods (93.4% and 86.2% classification accuracy, respectively).

READ FULL TEXT

page 1

page 2

page 3

page 4

research
07/27/2022

Point-McBert: A Multi-choice Self-supervised Framework for Point Cloud Pre-training

Masked language modeling (MLM) has become one of the most successful sel...
research
01/20/2022

CP-Net: Contour-Perturbed Reconstruction Network for Self-Supervised Point Cloud Learning

Self-supervised learning has not been fully explored for point cloud ana...
research
03/26/2022

Self-Supervised Point Cloud Representation Learning with Occlusion Auto-Encoder

Learning representations for point clouds is an important task in 3D com...
research
11/13/2022

Point-DAE: Denoising Autoencoders for Self-supervised Point Cloud Learning

Masked autoencoder has demonstrated its effectiveness in self-supervised...
research
12/10/2022

Complete-to-Partial 4D Distillation for Self-Supervised Point Cloud Sequence Representation Learning

Recent work on 4D point cloud sequences has attracted a lot of attention...
research
08/31/2023

CL-MAE: Curriculum-Learned Masked Autoencoders

Masked image modeling has been demonstrated as a powerful pretext task f...
research
01/12/2022

Local2Global: A distributed approach for scaling representation learning on graphs

We propose a decentralised "local2global"' approach to graph representat...

Please sign up or login with your details

Forgot password? Click here to reset