TM2D: Bimodality Driven 3D Dance Generation via Music-Text Integration

04/05/2023
by   Kehong Gong, et al.
0

We propose a novel task for generating 3D dance movements that simultaneously incorporate both text and music modalities. Unlike existing works that generate dance movements using a single modality such as music, our goal is to produce richer dance movements guided by the instructive information provided by the text. However, the lack of paired motion data with both music and text modalities limits the ability to generate dance movements that integrate both. To alleviate this challenge, we propose to utilize a 3D human motion VQ-VAE to project the motions of the two datasets into a latent space consisting of quantized vectors, which effectively mix the motion tokens from the two datasets with different distributions for training. Additionally, we propose a cross-modal transformer to integrate text instructions into motion generation architecture for generating 3D dance movements without degrading the performance of music-conditioned dance generation. To better evaluate the quality of the generated motion, we introduce two novel metrics, namely Motion Prediction Distance (MPD) and Freezing Score, to measure the coherence and freezing percentage of the generated motion. Extensive experiments show that our approach can generate realistic and coherent dance movements conditioned on both text and music while maintaining comparable performance with the two single modalities. Code will be available at: https://garfield-kh.github.io/TM2D/.

READ FULL TEXT

page 3

page 8

research
01/22/2023

Dance2MIDI: Dance-driven multi-instruments music generation

Dance-driven music generation aims to generate musical pieces conditione...
research
03/22/2023

Music-Driven Group Choreography

Music-driven choreography is a challenging problem with a wide variety o...
research
06/25/2021

Transflower: probabilistic autoregressive dance generation with multimodal attention

Dance requires skillful composition of complex movements that follow rhy...
research
09/03/2023

MAGMA: Music Aligned Generative Motion Autodecoder

Mapping music to dance is a challenging problem that requires spatial an...
research
05/12/2023

Unsupervised Melody-Guided Lyrics Generation

Automatic song writing is a topic of significant practical interest. How...
research
09/06/2023

MCM: Multi-condition Motion Synthesis Framework for Multi-scenario

The objective of the multi-condition human motion synthesis task is to i...
research
01/30/2023

DanceAnyWay: Synthesizing Mixed-Genre 3D Dance Movements Through Beat Disentanglement

We present DanceAnyWay, a hierarchical generative adversarial learning m...

Please sign up or login with your details

Forgot password? Click here to reset