LakhNES: Improving multi-instrumental music generation with cross-domain pre-training

07/10/2019
by   Chris Donahue, et al.
0

We are interested in the task of generating multi-instrumental music scores. The Transformer architecture has recently shown great promise for the task of piano score generation; here we adapt it to the multi-instrumental setting. Transformers are complex, high-dimensional language models which are capable of capturing long-term structure in sequence data, but require large amounts of data to fit. Their success on piano score generation is partially explained by the large volumes of symbolic data readily available for that domain. We leverage the recently-introduced NES-MDB dataset of four-instrument scores from an early video game sound synthesis chip (the NES), which we find to be well-suited to training with the Transformer architecture. To further improve the performance of our model, we propose a pre-training technique to leverage the information in a large collection of heterogeneous music, namely the Lakh MIDI dataset. Despite differences between the two corpora, we find that this transfer learning procedure improves both quantitative and qualitative performance for our primary task.

READ FULL TEXT
research
09/19/2023

MelodyGLM: Multi-task Pre-training for Symbolic Melody Generation

Pre-trained language models have achieved impressive results in various ...
research
06/01/2023

Transfer Learning for Underrepresented Music Generation

This paper investigates a combinational creativity approach to transfer ...
research
01/03/2023

Language Models are Drummers: Drum Composition with Natural Language Pre-Training

Automatic music generation with artificial intelligence typically requir...
research
11/21/2022

Exploring the Efficacy of Pre-trained Checkpoints in Text-to-Music Generation Task

Benefiting from large-scale datasets and pre-trained models, the field o...
research
06/09/2023

Reconstructing Human Expressiveness in Piano Performances with a Transformer Network

Capturing intricate and subtle variations in human expressiveness in mus...
research
07/21/2022

Unsupervised pre-training of graph transformers on patient population graphs

Pre-training has shown success in different areas of machine learning, s...
research
12/10/2019

Encoding Musical Style with Transformer Autoencoders

We consider the problem of learning high-level controls over the global ...

Please sign up or login with your details

Forgot password? Click here to reset