Visualizing and Understanding Self-attention based Music Tagging

11/11/2019
by   Minz Won, et al.
0

Recently, we proposed a self-attention based music tagging model. Different from most of the conventional deep architectures in music information retrieval, which use stacked 3x3 filters by treating music spectrograms as images, the proposed self-attention based model attempted to regard music as a temporal sequence of individual audio events. Not only the performance, but it could also facilitate better interpretability. In this paper, we mainly focus on visualizing and understanding the proposed self-attention based music tagging model.

READ FULL TEXT
research
06/12/2019

Toward Interpretable Music Tagging with Self-Attention

Self-attention is an attention mechanism that learns a representation by...
research
07/21/2021

Melody Structure Transfer Network: Generating Music with Separable Self-Attention

Symbolic music generation has attracted increasing attention, while most...
research
08/24/2022

Improved Zero-Shot Audio Tagging Classification with Patchout Spectrogram Transformers

Standard machine learning models for tagging and classifying acoustic si...
research
11/16/2019

Music theme recognition using CNN and self-attention

We present an efficient architecture to detect mood/themes in music trac...
research
02/18/2022

Deep-Learning Architectures for Multi-Pitch Estimation: Towards Reliable Evaluation

Extracting pitch information from music recordings is a challenging but ...
research
10/21/2020

AttendAffectNet: Self-Attention based Networks for Predicting Affective Responses from Movies

In this work, we propose different variants of the self-attention based ...
research
05/12/2021

Global Structure-Aware Drum Transcription Based on Self-Attention Mechanisms

This paper describes an automatic drum transcription (ADT) method that d...

Please sign up or login with your details

Forgot password? Click here to reset