About

Log in?

DTU users get better search results including licensed content and discounts on order fees.

Anyone can log in and get personalized features such as favorites, tags and feeds.

Log in as DTU user Log in as non-DTU user No thanks

DTU Findit

Conference paper

Modeling Temporal Structure in Music for Emotion Prediction using Pairwise Comparisons

In Proceedings of the 15th International Society for Music Information Retrieval Conference (ismir 2014) — 2014, pp. 319-324
From

Department of Applied Mathematics and Computer Science, Technical University of Denmark1

Cognitive Systems, Department of Applied Mathematics and Computer Science, Technical University of Denmark2

The temporal structure of music is essential for the cognitive processes related to the emotions expressed in music. However, such temporal information is often disregarded in typical Music Information Retrieval modeling tasks of predicting higher-level cognitive or semantic aspects of music such as emotions, genre, and similarity.

This paper addresses the specific hypothesis whether temporal information is essential for predicting expressed emotions in music, as a prototypical example of a cognitive aspect of music. We propose to test this hypothesis using a novel processing pipeline: 1) Extracting audio features for each track resulting in a multivariate ”feature time series”. 2) Using generative models to represent these time series (acquiring a complete track representation).

Specifically, we explore the Gaussian Mixture model, Vector Quantization, Autoregressive model, Markov and Hidden Markov models. 3) Utilizing the generative models in a discriminative setting by selecting the Probability Product Kernel as the natural kernel for all considered track representations. We evaluate the representations using a kernel based model specifically extended to support the robust two-alternative forced choice self-report paradigm, used for eliciting expressed emotions in music.

The methods are evaluated using two data sets and show increased predictive performance using temporal information, thus supporting the overall hypothesis.

Language: English
Publisher: International Society for Music Information Retrieval
Year: 2014
Pages: 319-324
Proceedings: 15th International Society for Music Information Retrieval Conference (ISMIR 2014)International Society for Music Information Retrieval Conference
Types: Conference paper
ORCIDs: Madsen, Jens , Jensen, Bjørn Sand and Larsen, Jan

DTU users get better search results including licensed content and discounts on order fees.

Log in as DTU user

Access

Analysis