Synthesizing Long-Term Human Motions with Diffusion Models via Coherent Sampling

08/03/2023
by   Zhao Yang, et al.
0

Text-to-motion generation has gained increasing attention, but most existing methods are limited to generating short-term motions that correspond to a single sentence describing a single action. However, when a text stream describes a sequence of continuous motions, the generated motions corresponding to each sentence may not be coherently linked. Existing long-term motion generation methods face two main issues. Firstly, they cannot directly generate coherent motions and require additional operations such as interpolation to process the generated actions. Secondly, they generate subsequent actions in an autoregressive manner without considering the influence of future actions on previous ones. To address these issues, we propose a novel approach that utilizes a past-conditioned diffusion model with two optional coherent sampling methods: Past Inpainting Sampling and Compositional Transition Sampling. Past Inpainting Sampling completes subsequent motions by treating previous motions as conditions, while Compositional Transition Sampling models the distribution of the transition as the composition of two adjacent motions guided by different text prompts. Our experimental results demonstrate that our proposed method is capable of generating compositional and coherent long-term 3D human motions controlled by a user-instructed long text stream. The code is available at \href{https://github.com/yangzhao1230/PCMDM}{https://github.com/yangzhao1230/PCMDM}.

READ FULL TEXT

page 1

page 4

page 8

research
12/12/2022

MultiAct: Long-Term 3D Human Motion Generation from Multiple Action Labels

We tackle the problem of generating long-term 3D human motion from multi...
research
08/23/2023

LongDanceDiff: Long-term Dance Generation with Conditional Diffusion Model

Dancing with music is always an essential human art form to express emot...
research
04/20/2023

SINC: Spatial Composition of 3D Human Motions for Simultaneous Action Generation

Our goal is to synthesize 3D human motions given textual inputs describi...
research
09/09/2022

TEACH: Temporal Action Composition for 3D Humans

Given a series of natural language descriptions, our task is to generate...
research
08/17/2020

Narrative Interpolation for Generating and Understanding Stories

We propose a method for controlled narrative/story generation where we a...
research
09/27/2022

NEURAL MARIONETTE: A Transformer-based Multi-action Human Motion Synthesis System

We present a neural network-based system for long-term, multi-action hum...
research
03/26/2021

Synthesis of Compositional Animations from Textual Descriptions

"How can we animate 3D-characters from a movie script or move robots by ...

Please sign up or login with your details

Forgot password? Click here to reset