Quantized GAN for Complex Music Generation from Dance Videos

04/01/2022
by   Ye Zhu, et al.
4

We present Dance2Music-GAN (D2M-GAN), a novel adversarial multi-modal framework that generates complex musical samples conditioned on dance videos. Our proposed framework takes dance video frames and human body motion as input, and learns to generate music samples that plausibly accompany the corresponding input. Unlike most existing conditional music generation works that generate specific types of mono-instrumental sounds using symbolic audio representations (e.g., MIDI), and that heavily rely on pre-defined musical synthesizers, in this work we generate dance music in complex styles (e.g., pop, breakdancing, etc.) by employing a Vector Quantized (VQ) audio representation, and leverage both its generality and the high abstraction capacity of its symbolic and continuous counterparts. By performing an extensive set of experiments on multiple datasets, and following a comprehensive evaluation protocol, we assess the generative quality of our approach against several alternatives. The quantitative results, which measure the music consistency, beats correspondence, and music diversity, clearly demonstrate the effectiveness of our proposed method. Last but not least, we curate a challenging dance-music dataset of in-the-wild TikTok videos, which we use to further demonstrate the efficacy of our approach in real-world applications - and which we hope to serve as a starting point for relevant future research.

READ FULL TEXT

page 4

page 6

research
05/11/2023

V2Meow: Meowing to the Visual Beat via Music Generation

Generating high quality music that complements the visual content of a v...
research
07/21/2020

Foley Music: Learning to Generate Music from Videos

In this paper, we introduce Foley Music, a system that can synthesize pl...
research
06/23/2020

Audeo: Audio Generation for a Silent Performance Video

We present a novel system that gets as an input video frames of a musici...
research
03/22/2023

Music-Driven Group Choreography

Music-driven choreography is a challenging problem with a wide variety o...
research
08/03/2023

MusicLDM: Enhancing Novelty in Text-to-Music Generation Using Beat-Synchronous Mixup Strategies

Diffusion models have shown promising results in cross-modal generation ...
research
06/06/2023

Dance Generation by Sound Symbolic Words

This study introduces a novel approach to generate dance motions using o...
research
06/23/2017

Toward Faultless Content-Based Playlists Generation for Instrumentals

This study deals with content-based musical playlists generation focused...

Please sign up or login with your details

Forgot password? Click here to reset