Exploiting Synchronized Lyrics And Vocal Features For Music Emotion Detection

01/15/2019
by   Loreto Parisi, et al.
0

One of the key points in music recommendation is authoring engaging playlists according to sentiment and emotions. While previous works were mostly based on audio for music discovery and playlists generation, we take advantage of our synchronized lyrics dataset to combine text representations and music features in a novel way; we therefore introduce the Synchronized Lyrics Emotion Dataset. Unlike other approaches that randomly exploited the audio samples and the whole text, our data is split according to the temporal information provided by the synchronization between lyrics and audio. This work shows a comparison between text-based and audio-based deep learning classification models using different techniques from Natural Language Processing and Music Information Retrieval domains. From the experiments on audio we conclude that using vocals only, instead of the whole audio data improves the overall performances of the audio classifier. In the lyrics experiments we exploit the state-of-the-art word representations applied to the main Deep Learning architectures available in literature. In our benchmarks the results show how the Bilinear LSTM classifier with Attention based on fastText word embedding performs better than the CNN applied on audio.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
04/13/2021

Comparison and Analysis of Deep Audio Embeddings for Music Emotion Recognition

Emotion is a complicated notion present in music that is hard to capture...
research
01/20/2022

Kinit Classification in Ethiopian Chants, Azmaris and Modern Music: A New Dataset and CNN Benchmark

In this paper, we create EMIR, the first-ever Music Information Retrieva...
research
01/12/2023

Rock Guitar Tablature Generation via Natural Language Processing

Deep learning has recently empowered and democratized generative modelin...
research
08/31/2023

Sequential Pitch Distributions for Raga Detection

Raga is a fundamental melodic concept in Indian Art Music (IAM). It is c...
research
02/26/2023

Multi-Modality in Music: Predicting Emotion in Music from High-Level Audio Features and Lyrics

This paper aims to test whether a multi-modal approach for music emotion...
research
07/18/2021

Residual Attention Based Network for Automatic Classification of Phonation Modes

Phonation mode is an essential characteristic of singing style as well a...
research
05/23/2023

When the Music Stops: Tip-of-the-Tongue Retrieval for Music

We present a study of Tip-of-the-tongue (ToT) retrieval for music, where...

Please sign up or login with your details

Forgot password? Click here to reset