Predictive modeling of expressed emotions in music using pairwise comparisons

Madsen, J., Sand Jensen, B. and Larsen, J. (2013) Predictive modeling of expressed emotions in music using pairwise comparisons. Lecture Notes in Computer Science, 7900, pp. 253-277. (doi: 10.1007/978-3-642-41248-6_14)

Full text not currently available from Enlighten.


We introduce a two-alternative forced-choice (2AFC) experimental paradigm to quantify expressed emotions in music using the arousal and valence (AV) dimensions. A wide range of well-known audio features are investigated for predicting the expressed emotions in music using learning curves and essential baselines. We furthermore investigate the scalability issues of using 2AFC in quantifying emotions expressed in music on large-scale music databases. The possibility of dividing the annotation task between multiple individuals, while pooling individuals’ comparisons is investigated by looking at the subjective differences of ranking emotion in the AV space. We find this to be problematic due to the large variation in subjects’ rankings of excerpts. Finally, solving scalability issues by reducing the number of pairwise comparisons is analyzed. We compare two active learning schemes to selecting comparisons at random by using learning curves. We show that a suitable predictive model of expressed valence in music can be achieved from only 15% of the total number of comparisons when using the Expected Value of Information (EVOI) active learning scheme. For the arousal dimension we require 9% of the total number of comparisons.

Item Type:Articles
Glasgow Author(s) Enlighten ID:Jensen, Dr Bjorn
Authors: Madsen, J., Sand Jensen, B., and Larsen, J.
College/School:College of Science and Engineering > School of Computing Science
Journal Name:Lecture Notes in Computer Science

University Staff: Request a correction | Enlighten Editors: Update this record