User Specific Adaptation in Automatic Transcription of Vocalised Percussion

11/06/2018
by   António Ramires, et al.
0

The goal of this work is to develop an application that enables music producers to use their voice to create drum patterns when composing in Digital Audio Workstations (DAWs). An easy-to-use and user-oriented system capable of automatically transcribing vocalisations of percussion sounds, called LVT - Live Vocalised Transcription, is presented. LVT is developed as a Max for Live device which follows the `segment-and-classify' methodology for drum transcription, and includes three modules: i) an onset detector to segment events in time; ii) a module that extracts relevant features from the audio content; and iii) a machine-learning component that implements the k-Nearest Neighbours (kNN) algorithm for the classification of vocalised drum timbres. Due to the wide differences in vocalisations from distinct users for the same drum sound, a user-specific approach to vocalised transcription is proposed. In this perspective, a given end-user trains the algorithm with their own vocalisations for each drum sound before inputting their desired pattern into the DAW. The user adaption is achieved via a new Max external which implements Sequential Forward Selection (SFS) for choosing the most relevant features for a given set of input drum sounds.

READ FULL TEXT
research
04/26/2021

Identifying Actions for Sound Event Classification

In Psychology, actions are paramount for humans to perceive and separate...
research
06/03/2021

LyricJam: A system for generating lyrics for live instrumental music

We describe a real-time system that receives a live audio stream from a ...
research
11/08/2019

Automatic Identification of Traditional Colombian Music Genres based on Audio Content Analysis and Machine Learning Technique

Colombia has a diversity of genres in traditional music, which allows to...
research
10/27/2022

LyricJam Sonic: A Generative System for Real-Time Composition and Musical Improvisation

Electronic music artists and sound designers have unique workflow practi...
research
07/10/2023

EchoVest: Real-Time Sound Classification and Depth Perception Expressed through Transcutaneous Electrical Nerve Stimulation

Over 1.5 billion people worldwide live with hearing impairment. Despite ...
research
02/02/2015

Unsupervised Incremental Learning and Prediction of Music Signals

A system is presented that segments, clusters and predicts musical audio...
research
01/21/2019

Learning sound representations using trainable COPE feature extractors

Sound analysis research has mainly been focused on speech and music proc...

Please sign up or login with your details

Forgot password? Click here to reset