Compound Word Transformer: Learning to Compose Full-Song Music over Dynamic Directed Hypergraphs

01/07/2021
by   Wen-Yi Hsiao, et al.
0

To apply neural sequence models such as the Transformers to music generation tasks, one has to represent a piece of music by a sequence of tokens drawn from a finite set of pre-defined vocabulary. Such a vocabulary usually involves tokens of various types. For example, to describe a musical note, one needs separate tokens to indicate the note's pitch, duration, velocity (dynamics), and placement (onset time) along the time grid. While different types of tokens may possess different properties, existing models usually treat them equally, in the same way as modeling words in natural languages. In this paper, we present a conceptually different approach that explicitly takes into account the type of the tokens, such as note types and metric types. And, we propose a new Transformer decoder architecture that uses different feed-forward heads to model tokens of different types. With an expansion-compression trick, we convert a piece of music to a sequence of compound words by grouping neighboring tokens, greatly reducing the length of the token sequences. We show that the resulting model can be viewed as a learner over dynamic directed hypergraphs. And, we employ it to learn to compose expressive Pop piano music of full-song length (involving up to 10K individual tokens per song), both conditionally and unconditionally. Our experiment shows that, compared to state-of-the-art models, the proposed model converges 5–10 times faster at training (i.e., within a day on a single GPU with 11 GB memory), and with comparable quality in the generated music.

READ FULL TEXT
research
01/06/2023

Multi-Genre Music Transformer – Composing Full Length Musical Piece

In the task of generating music, the art factor plays a big role and is ...
research
04/09/2021

Heaps' Law and Vocabulary Richness in the History of Classical Music Harmony

Music is a fundamental human construct, and harmony provides the buildin...
research
11/01/2021

Learning To Generate Piano Music With Sustain Pedals

Recent years have witnessed a growing interest in research related to th...
research
05/22/2023

FIT: Far-reaching Interleaved Transformers

We present FIT: a transformer-based architecture with efficient self-att...
research
07/10/2023

VampNet: Music Generation via Masked Acoustic Token Modeling

We introduce VampNet, a masked acoustic token modeling approach to music...
research
07/21/2020

Word Representation for Rhythms

This paper proposes a word representation strategy for rhythm patterns. ...
research
01/31/2023

An Comparative Analysis of Different Pitch and Metrical Grid Encoding Methods in the Task of Sequential Music Generation

Pitch and meter are two fundamental music features for symbolic music ge...

Please sign up or login with your details

Forgot password? Click here to reset