MuseMorphose: Full-Song and Fine-Grained Music Style Transfer with One Transformer VAE

05/10/2021
by   Shih-Lun Wu, et al.
0

Transformers and variational autoencoders (VAE) have been extensively employed for symbolic (e.g., MIDI) domain music generation. While the former boast an impressive capability in modeling long sequences, the latter allow users to willingly exert control over different parts (e.g., bars) of the music to be generated. In this paper, we are interested in bringing the two together to construct a single model that exhibits both strengths. The task is split into two steps. First, we equip Transformer decoders with the ability to accept segment-level, time-varying conditions during sequence generation. Subsequently, we combine the developed and tested in-attention decoder with a Transformer encoder, and train the resulting MuseMorphose model with the VAE objective to achieve style transfer of long musical pieces, in which users can specify musical attributes including rhythmic intensity and polyphony (i.e., harmonic fullness) they desire, down to the bar level. Experiments show that MuseMorphose outperforms recurrent neural network (RNN) based baselines on numerous widely-used metrics for style transfer tasks.

READ FULL TEXT

page 23

page 24

page 25

research
09/20/2018

MIDI-VAE: Modeling Dynamics and Instrumentation of Music with Applications to Style Transfer

We introduce MIDI-VAE, a neural network model based on Variational Autoe...
research
05/17/2022

The Power of Reuse: A Multi-Scale Transformer Model for Structural Dynamic Segmentation in Symbolic Music Generation

Symbolic Music Generation relies on the contextual representation capabi...
research
12/10/2019

Encoding Musical Style with Transformer Autoencoders

We consider the problem of learning high-level controls over the global ...
research
02/10/2021

Self-Supervised VQ-VAE For One-Shot Music Style Transfer

Neural style transfer, allowing to apply the artistic style of one image...
research
07/30/2021

DadaGP: A Dataset of Tokenized GuitarPro Songs for Sequence Models

Originating in the Renaissance and burgeoning in the digital era, tablat...
research
05/29/2019

Revision in Continuous Space: Fine-Grained Control of Text Style Transfer

Typical methods for unsupervised text style transfer often rely on two k...
research
06/21/2019

Query-based Deep Improvisation

In this paper we explore techniques for generating new music using a Var...

Please sign up or login with your details

Forgot password? Click here to reset