AssemblyHands: Towards Egocentric Activity Understanding via 3D Hand Pose Estimation

04/24/2023
by   Takehiko Ohkawa, et al.
0

We present AssemblyHands, a large-scale benchmark dataset with accurate 3D hand pose annotations, to facilitate the study of egocentric activities with challenging hand-object interactions. The dataset includes synchronized egocentric and exocentric images sampled from the recent Assembly101 dataset, in which participants assemble and disassemble take-apart toys. To obtain high-quality 3D hand pose annotations for the egocentric images, we develop an efficient pipeline, where we use an initial set of manual annotations to train a model to automatically annotate a much larger dataset. Our annotation model uses multi-view feature fusion and an iterative refinement scheme, and achieves an average keypoint error of 4.20 mm, which is 85 original annotations in Assembly101. AssemblyHands provides 3.0M annotated images, including 490K egocentric images, making it the largest existing benchmark dataset for egocentric 3D hand pose estimation. Using this data, we develop a strong single-view baseline of 3D hand pose estimation from egocentric images. Furthermore, we design a novel action classification task to evaluate predicted 3D hand poses. Our study shows that having higher-quality hand poses directly improves the ability to recognize actions.

READ FULL TEXT

page 2

page 5

page 7

page 11

page 12

research
12/06/2020

MVHM: A Large-Scale Multi-View Hand Mesh Benchmark for Accurate 3D Hand Pose Estimation

Estimating 3D hand poses from a single RGB image is challenging because ...
research
03/28/2022

Assembly101: A Large-Scale Multi-View Video Dataset for Understanding Procedural Activities

Assembly101 is a new procedural activity dataset featuring 4321 videos o...
research
06/10/2022

Ego2HandsPose: A Dataset for Egocentric Two-hand 3D Global Pose Estimation

Color-based two-hand 3D pose estimation in the global coordinate system ...
research
07/02/2021

HO-3D_v3: Improving the Accuracy of Hand-Object Annotations of the HO-3D Dataset

HO-3D is a dataset providing image sequences of various hand-object inte...
research
07/06/2023

Self-supervised Optimization of Hand Pose Estimation using Anatomical Features and Iterative Learning

Manual assembly workers face increasing complexity in their work. Human-...
research
06/05/2022

Efficient Annotation and Learning for 3D Hand Pose Estimation: A Survey

In this survey, we present comprehensive analysis of 3D hand pose estima...
research
07/12/2017

Large-scale Multiview 3D Hand Pose Dataset

Accurate hand pose estimation at joint level has several uses on human-r...

Please sign up or login with your details

Forgot password? Click here to reset