DocumentCode
697857
Title
Music genre classification via sparse representations of auditory temporal modulations
Author
Panagakis, Yannis ; Kotropoulos, Constantine ; Arce, Gonzalo R.
Author_Institution
Dept. of Inf., Aristotle Univ. of Thessaloniki, Thessaloniki, Greece
fYear
2009
fDate
24-28 Aug. 2009
Firstpage
1
Lastpage
5
Abstract
A robust music genre classification framework is proposed that combines the rich, psycho-physiologically grounded properties of slow temporal modulations of music recordings and the power of sparse representation-based classifiers. Linear subspace dimensionality reduction techniques are shown to play a crucial role within the framework under study. The proposed method yields a music genre classification accuracy of 91% and 93.56% on the GTZAN and the ISMIR2004 Genre dataset, respectively. Both accuracies outperform any reported accuracy ever obtained by state of the art music genre classification algorithms in the aforementioned datasets.
Keywords
audio recording; audio signal processing; compressed sensing; modulation; music; signal classification; signal representation; GTZAN; ISMIR2004 genre dataset; auditory temporal modulations; linear subspace dimensionality reduction techniques; music genre classification algorithms; music recordings; sparse representation-based classifiers; Accuracy; Dictionaries; Feature extraction; Modulation; Music; Principal component analysis; Training;
fLanguage
English
Publisher
ieee
Conference_Titel
Signal Processing Conference, 2009 17th European
Conference_Location
Glasgow
Print_ISBN
978-161-7388-76-7
Type
conf
Filename
7077429
Link To Document