DocumentCode :
697857
Title :
Music genre classification via sparse representations of auditory temporal modulations
Author :
Panagakis, Yannis ; Kotropoulos, Constantine ; Arce, Gonzalo R.
Author_Institution :
Dept. of Inf., Aristotle Univ. of Thessaloniki, Thessaloniki, Greece
fYear :
2009
fDate :
24-28 Aug. 2009
Firstpage :
1
Lastpage :
5
Abstract :
A robust music genre classification framework is proposed that combines the rich, psycho-physiologically grounded properties of slow temporal modulations of music recordings and the power of sparse representation-based classifiers. Linear subspace dimensionality reduction techniques are shown to play a crucial role within the framework under study. The proposed method yields a music genre classification accuracy of 91% and 93.56% on the GTZAN and the ISMIR2004 Genre dataset, respectively. Both accuracies outperform any reported accuracy ever obtained by state of the art music genre classification algorithms in the aforementioned datasets.
Keywords :
audio recording; audio signal processing; compressed sensing; modulation; music; signal classification; signal representation; GTZAN; ISMIR2004 genre dataset; auditory temporal modulations; linear subspace dimensionality reduction techniques; music genre classification algorithms; music recordings; sparse representation-based classifiers; Accuracy; Dictionaries; Feature extraction; Modulation; Music; Principal component analysis; Training;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Signal Processing Conference, 2009 17th European
Conference_Location :
Glasgow
Print_ISBN :
978-161-7388-76-7
Type :
conf
Filename :
7077429
Link To Document :
بازگشت