Analysis of Hand Segmentation in the Wild

03/08/2018
by   Aisha Urooj Khan, et al.
0

A large number of works in egocentric vision have concentrated on action and object recognition. Detection and segmentation of hands in first person videos, however, has less been explored. For many applications in this domain, it is necessary to accurately segment not only hands of the camera wearer but also the hands of others with whom he is interacting. Here, we take an in-depth look at the hand segmentation problem. First, we evaluate the performance of the state of the art hand segmentation methods, off the shelf and finetuned, on existing datasets. Second, we finetune RefineNet, a leading semantic segmentation method, for hand segmentation and find that it does much better than best contenders. Third, we contribute by collecting two new datasets including a) EgoYouTubeHands dataset which includes egocentric videos containing hands in the wild, and b) HandOverFace dataset to analyze the performance of our models in presence of similar appearance occlusions. Fourth, we investigate whether conditional random fields can be helpful to refine hand segmentations produced by our model. Fifth, we train a CNN for hand-based activity recognition and achieve higher activity recognition accuracy when the trained CNN used hand maps produced by finetuned RefineNet model. Finally, we annotate a subset of the EgoHands dataset for fine-level activity recognition and show that just looking at a single hand pose, we can achieve 58.6 recognition accuracy where chance level is 12.5

READ FULL TEXT

page 3

page 5

page 6

page 7

page 8

research
07/21/2020

Creating a Large-scale Synthetic Dataset for Human Activity Recognition

Creating and labelling datasets of videos for use in training Human Acti...
research
06/16/2018

Object Level Visual Reasoning in Videos

Human activity recognition is typically addressed by training models to ...
research
08/07/2022

Fine-Grained Egocentric Hand-Object Segmentation: Dataset, Model, and Applications

Egocentric videos offer fine-grained information for high-fidelity model...
research
03/11/2019

A Hybrid Framework for Action Recognition in Low-Quality Video Sequences

Vision-based activity recognition is essential for security, monitoring ...
research
03/29/2019

DenseAttentionSeg: Segment Hands from Interacted Objects Using Depth Input

We propose a real-time DNN-based technique to segment hand and object of...
research
02/28/2020

Hand-Priming in Object Localization for Assistive Egocentric Vision

Egocentric vision holds great promises for increasing access to visual i...
research
08/16/2022

Multi-level Contrast Network for Wearables-based Joint Activity Segmentation and Recognition

Human activity recognition (HAR) with wearables is promising research th...

Please sign up or login with your details

Forgot password? Click here to reset