DocumentCode
730153
Title
A histogram density modeling approach to music emotion recognition
Author
Ju-Chiang Wang ; Hsin-Min Wang ; Lanckriet, Gert
Author_Institution
Dept. of Electr. & Comput. Eng., UC San Diego, San Diego, CA, USA
fYear
2015
fDate
19-24 April 2015
Firstpage
698
Lastpage
702
Abstract
Music emotion recognition is concerned with developing predictive models that comprehend the affective content of musical signals. Recently, a growing number of attempts has been made to model the music emotion as a probability distribution in the valence-arousal (VA) space to better account for the subjectivity. In this paper, we present a novel histogram density modeling approach that models the emotion distribution by a 2-D histogram over the quantized VA space and learns a set of latent histograms to predict the emotion probability density of a song from audio. The proposed model is free from parametric distribution assumptions over the VA space, easy to implement, and extremely fast to train. We also extend our model to deal with the temporal dynamics of time-varying emotion labels. Comprehensive performance study on two larger-scale datasets demonstrates that our approach achieves comparable performance to the state-of-the-art ones, but with much better training and testing efficiency.
Keywords
acoustic signal processing; emotion recognition; music; statistical distributions; 2D histogram; VA space; emotion distribution; emotion probability density prediction; histogram density modeling approach; music emotion recognition; probability distribution; time varying emotion label temporal dynamics; valence-arousal space; Histograms; Unified modeling language; Affective computing; emotion tracking; music information retrieval; subjectivity; temporal dynamics;
fLanguage
English
Publisher
ieee
Conference_Titel
Acoustics, Speech and Signal Processing (ICASSP), 2015 IEEE International Conference on
Conference_Location
South Brisbane, QLD
Type
conf
DOI
10.1109/ICASSP.2015.7178059
Filename
7178059
Link To Document