Sign Language Recognition via Skeleton-Aware Multi-Model Ensemble

10/12/2021
by   Songyao Jiang, et al.
0

Sign language is commonly used by deaf or mute people to communicate but requires extensive effort to master. It is usually performed with the fast yet delicate movement of hand gestures, body posture, and even facial expressions. Current Sign Language Recognition (SLR) methods usually extract features via deep neural networks and suffer overfitting due to limited and noisy data. Recently, skeleton-based action recognition has attracted increasing attention due to its subject-invariant and background-invariant nature, whereas skeleton-based SLR is still under exploration due to the lack of hand annotations. Some researchers have tried to use off-line hand pose trackers to obtain hand keypoints and aid in recognizing sign language via recurrent neural networks. Nevertheless, none of them outperforms RGB-based approaches yet. To this end, we propose a novel Skeleton Aware Multi-modal Framework with a Global Ensemble Model (GEM) for isolated SLR (SAM-SLR-v2) to learn and fuse multi-modal feature representations towards a higher recognition rate. Specifically, we propose a Sign Language Graph Convolution Network (SL-GCN) to model the embedded dynamics of skeleton keypoints and a Separable Spatial-Temporal Convolution Network (SSTCN) to exploit skeleton features. The skeleton-based predictions are fused with other RGB and depth based modalities by the proposed late-fusion GEM to provide global information and make a faithful SLR prediction. Experiments on three isolated SLR datasets demonstrate that our proposed SAM-SLR-v2 framework is exceedingly effective and achieves state-of-the-art performance with significant margins. Our code will be available at https://github.com/jackyjsy/SAM-SLR-v2

READ FULL TEXT

page 1

page 3

page 4

page 5

page 10

research
03/16/2021

Skeleton Based Sign Language Recognition Using Whole-body Keypoints

Sign language is a visual language that is used by deaf or speech impair...
research
04/10/2023

Isolated Sign Language Recognition based on Tree Structure Skeleton Images

Sign Language Recognition (SLR) systems aim to be embedded in video stre...
research
12/25/2022

StepNet: Spatial-temporal Part-aware Network for Sign Language Recognition

Sign language recognition (SLR) aims to overcome the communication barri...
research
09/21/2023

SlowFast Network for Continuous Sign Language Recognition

The objective of this work is the effective extraction of spatial and dy...
research
12/21/2022

SLGTformer: An Attention-Based Approach to Sign Language Recognition

Sign language is the preferred method of communication of deaf or mute p...
research
02/03/2022

Exploring Sub-skeleton Trajectories for Interpretable Recognition of Sign Language

Recent advances in tracking sensors and pose estimation software enable ...
research
06/07/2019

Recognizing American Sign Language Manual Signs from RGB-D Videos

In this paper, we propose a 3D Convolutional Neural Network (3DCNN) base...

Please sign up or login with your details

Forgot password? Click here to reset