Can Everybody Sign Now? Exploring Sign Language Video Generation from 2D Poses

12/20/2020
by   Lucas Ventura, et al.
10

Recent work have addressed the generation of human poses represented by 2D/3D coordinates of human joints for sign language. We use the state of the art in Deep Learning for motion transfer and evaluate them on How2Sign, an American Sign Language dataset, to generate videos of signers performing sign language given a 2D pose skeleton. We evaluate the generated videos quantitatively and qualitatively showing that the current models are not enough to generated adequate videos for Sign Language due to lack of detail in hands.

READ FULL TEXT
research
01/10/2023

Learning from What is Already Out There: Few-shot Sign Language Recognition with Online Dictionaries

Today's sign language recognition models require large training corpora ...
research
06/24/2021

Towards Automatic Speech to Sign Language Generation

We aim to solve the highly challenging task of generating continuous sig...
research
05/06/2021

Pose-Guided Sign Language Video GAN with Dynamic Lambda

We propose a novel approach for the synthesis of sign language videos us...
research
10/01/2021

Phonology Recognition in American Sign Language

Inspired by recent developments in natural language processing, we propo...
research
07/08/2022

Jointly Harnessing Prior Structures and Temporal Consistency for Sign Language Video Generation

Sign language is the window for people differently-abled to express thei...
research
08/18/2023

Human Part-wise 3D Motion Context Learning for Sign Language Recognition

In this paper, we propose P3D, the human part-wise motion context learni...
research
05/26/2023

Motion-Based Sign Language Video Summarization using Curvature and Torsion

An interesting problem in many video-based applications is the generatio...

Please sign up or login with your details

Forgot password? Click here to reset