K-Order Graph-oriented Transformer with GraAttention for 3D Pose and Shape Estimation

08/24/2022
by   Weixi Zhao, et al.
0

We propose a novel attention-based 2D-to-3D pose estimation network for graph-structured data, named KOG-Transformer, and a 3D pose-to-shape estimation network for hand data, named GASE-Net. Previous 3D pose estimation methods have focused on various modifications to the graph convolution kernel, such as abandoning weight sharing or increasing the receptive field. Some of these methods employ attention-based non-local modules as auxiliary modules. In order to better model the relationship between nodes in graph-structured data and fuse the information of different neighbor nodes in a differentiated way, we make targeted modifications to the attention module and propose two modules designed for graph-structured data, graph relative positional encoding multi-head self-attention (GR-MSA) and K-order graph-oriented multi-head self-attention (KOG-MSA). By stacking GR-MSA and KOG-MSA, we propose a novel network KOG-Transformer for 2D-to-3D pose estimation. Furthermore, we propose a network for shape estimation on hand data, called GraAttention shape estimation network (GASE-Net), which takes a 3D pose as input and gradually models the shape of the hand from sparse to dense. We have empirically shown the superiority of KOG-Transformer through extensive experiments. Experimental results show that KOG-Transformer significantly outperforms the previous state-of-the-art methods on the benchmark dataset Human3.6M. We evaluate the effect of GASE-Net on two public available hand datasets, ObMan and InterHand2.6M. GASE-Net can predict the corresponding shape for input pose with strong generalization ability.

READ FULL TEXT
research
09/17/2021

GraFormer: Graph Convolution Transformer for 3D Pose Estimation

Exploiting relations among 2D joints plays a crucial role yet remains se...
research
06/01/2022

Learning Sequential Contexts using Transformer for 3D Hand Pose Estimation

3D hand pose estimation (HPE) is the process of locating the joints of t...
research
02/15/2023

Pose-Oriented Transformer with Uncertainty-Guided Refinement for 2D-to-3D Human Pose Estimation

There has been a recent surge of interest in introducing transformers to...
research
03/23/2023

TransPoser: Transformer as an Optimizer for Joint Object Shape and Pose Estimation

We propose a novel method for joint estimation of shape and pose of rigi...
research
04/07/2023

A2J-Transformer: Anchor-to-Joint Transformer Network for 3D Interacting Hand Pose Estimation from a Single RGB Image

3D interacting hand pose estimation from a single RGB image is a challen...
research
03/09/2023

Deformer: Dynamic Fusion Transformer for Robust Hand Pose Estimation

Accurately estimating 3D hand pose is crucial for understanding how huma...
research
02/07/2022

HeadPosr: End-to-end Trainable Head Pose Estimation using Transformer Encoders

In this paper, HeadPosr is proposed to predict the head poses using a si...

Please sign up or login with your details

Forgot password? Click here to reset