Learning source-aware representations of music in a discrete latent space

11/26/2021
by   Jinsung Kim, et al.
6

In recent years, neural network based methods have been proposed as a method that cangenerate representations from music, but they are not human readable and hardly analyzable oreditable by a human. To address this issue, we propose a novel method to learn source-awarelatent representations of music through Vector-Quantized Variational Auto-Encoder(VQ-VAE).We train our VQ-VAE to encode an input mixture into a tensor of integers in a discrete latentspace, and design them to have a decomposed structure which allows humans to manipulatethe latent vector in a source-aware manner. This paper also shows that we can generate basslines by estimating latent vectors in a discrete space.

READ FULL TEXT
research
04/03/2018

Hyperspherical Variational Auto-Encoders

The Variational Auto-Encoder (VAE) is one of the most used unsupervised ...
research
08/19/2019

Audio query-based music source separation

In recent years, music source separation has been one of the most intens...
research
06/09/2019

Deep Music Analogy Via Latent Representation Disentanglement

Analogy is a key solution to automated music generation, featured by its...
research
02/12/2023

Vector Quantized Wasserstein Auto-Encoder

Learning deep discrete latent presentations offers a promise of better s...
research
05/18/2020

Robust Training of Vector Quantized Bottleneck Models

In this paper we demonstrate methods for reliable and efficient training...
research
04/11/2020

Attribute-based Regularization of VAE Latent Spaces

Selective manipulation of data attributes using deep generative models i...
research
04/21/2020

Vector Quantized Contrastive Predictive Coding for Template-based Music Generation

In this work, we propose a flexible method for generating variations of ...

Please sign up or login with your details

Forgot password? Click here to reset