TEMOS: Generating diverse human motions from textual descriptions

04/25/2022
by   Mathis Petrovich, et al.
0

We address the problem of generating diverse 3D human motions from textual descriptions. This challenging task requires joint modeling of both modalities: understanding and extracting useful human-centric information from the text, and then generating plausible and realistic sequences of human poses. In contrast to most previous work which focuses on generating a single, deterministic, motion from a textual description, we design a variational approach that can produce multiple diverse human motions. We propose TEMOS, a text-conditioned generative model leveraging variational autoencoder (VAE) training with human motion data, in combination with a text encoder that produces distribution parameters compatible with the VAE latent space. We show that TEMOS framework can produce both skeleton-based animations as in prior work, as well more expressive SMPL body motions. We evaluate our approach on the KIT Motion-Language benchmark and, despite being relatively straightforward, demonstrate significant improvements over the state of the art. Code and models are available on our project page.

READ FULL TEXT

page 11

page 14

research
12/08/2022

Executing your Commands via Motion Diffusion in Latent Space

We study a challenging task, conditional human motion generation, which ...
research
01/15/2023

T2M-GPT: Generating Human Motion from Textual Descriptions with Discrete Representations

In this work, we investigate a simple and must-known conditional generat...
research
11/25/2022

PaCMO: Partner Dependent Human Motion Generation in Dyadic Human Activity using Neural Operators

We address the problem of generating 3D human motions in dyadic activiti...
research
04/04/2022

HiT-DVAE: Human Motion Generation via Hierarchical Transformer Dynamical VAE

Studies on the automatic processing of 3D human pose data have flourishe...
research
08/28/2023

Priority-Centric Human Motion Generation in Discrete Latent Space

Text-to-motion generation is a formidable task, aiming to produce human ...
research
07/04/2022

TM2T: Stochastic and Tokenized Modeling for the Reciprocal Generation of 3D Human Motions and Texts

Inspired by the strong ties between vision and language, the two intimat...
research
12/16/2021

The Wanderings of Odysseus in 3D Scenes

Our goal is to populate digital environments, in which the digital human...

Please sign up or login with your details

Forgot password? Click here to reset