Multi-Modal Zero-Shot Sign Language Recognition

09/02/2021
by   Razieh Rastgoo, et al.
10

Zero-Shot Learning (ZSL) has rapidly advanced in recent years. Towards overcoming the annotation bottleneck in the Sign Language Recognition (SLR), we explore the idea of Zero-Shot Sign Language Recognition (ZS-SLR) with no annotated visual examples, by leveraging their textual descriptions. In this way, we propose a multi-modal Zero-Shot Sign Language Recognition (ZS-SLR) model harnessing from the complementary capabilities of deep features fused with the skeleton-based ones. A Transformer-based model along with a C3D model is used for hand detection and deep features extraction, respectively. To make a trade-off between the dimensionality of the skeletonbased and deep features, we use an Auto-Encoder (AE) on top of the Long Short Term Memory (LSTM) network. Finally, a semantic space is used to map the visual features to the lingual embedding of the class labels, achieved via the Bidirectional Encoder Representations from Transformers (BERT) model. Results on four large-scale datasets, RKS-PERSIANSIGN, First-Person, ASLVID, and isoGD, show the superiority of the proposed model compared to state-of-the-art alternatives in ZS-SLR.

READ FULL TEXT

page 5

page 9

research
08/23/2021

ZS-SLR: Zero-Shot Sign Language Recognition from RGB-D Videos

Sign Language Recognition (SLR) is a challenging research area in comput...
research
01/15/2022

Towards Zero-shot Sign Language Recognition

This paper tackles the problem of zero-shot sign language recognition (Z...
research
07/24/2019

Zero-Shot Sign Language Recognition: Can Textual Data Uncover Sign Languages?

We introduce the problem of zero-shot sign language recognition (ZSSLR),...
research
01/22/2020

Zero-Shot Activity Recognition with Videos

In this paper, we examined the zero-shot activity recognition task with ...
research
04/05/2022

A Transformer-Based Contrastive Learning Approach for Few-Shot Sign Language Recognition

Sign language recognition from sequences of monocular images or 2D poses...
research
09/22/2020

Visual Methods for Sign Language Recognition: A Modality-Based Review

Sign language visual recognition from continuous multi-modal streams is ...
research
08/03/2020

AUTSL: A Large Scale Multi-modal Turkish Sign Language Dataset and Baseline Methods

Sign language recognition is a challenging problem where signs are ident...

Please sign up or login with your details

Forgot password? Click here to reset