Author(s):
Panda, Renato ; Paiva, Rui Pedro
Date: 2011
Persistent ID: https://hdl.handle.net/10316/95172
Origin: Estudo Geral - Universidade de Coimbra
Project/scholarship:
info:eu-repo/grantAgreement/FCT/5876-PPCDTI/102185/PT
;
Subject(s): Mood tracking; Music emotion recognition; Regression; Thayer
Description
In this paper we propose a solution for automatic mood tracking in audio music, based on supervised learning and classification. To this end, various music clips with a duration of 25 seconds, previously annotated with arousal and valence (AV) values, were used to train several models. These models were used to predict quadrants of the Thayer’s taxonomy and AV values, of small segments from full songs, revealing the mood changes over time. The system accuracy was measured by calculating the matching ratio between predicted results and full song annotations performed by volunteers. Different combinations of audio features, frameworks and other parameters were tested, resulting in an accuracy of 56.3% and showing there is still much room for improvement.
This work was supported by the MOODetector project (PTDC/EIA-EIA/102185/2008), financed by the Fundação para Ciência e Tecnologia - Portugal.