• DocumentCode
    697857
  • Title

    Music genre classification via sparse representations of auditory temporal modulations

  • Author

    Panagakis, Yannis ; Kotropoulos, Constantine ; Arce, Gonzalo R.

  • Author_Institution
    Dept. of Inf., Aristotle Univ. of Thessaloniki, Thessaloniki, Greece
  • fYear
    2009
  • fDate
    24-28 Aug. 2009
  • Firstpage
    1
  • Lastpage
    5
  • Abstract
    A robust music genre classification framework is proposed that combines the rich, psycho-physiologically grounded properties of slow temporal modulations of music recordings and the power of sparse representation-based classifiers. Linear subspace dimensionality reduction techniques are shown to play a crucial role within the framework under study. The proposed method yields a music genre classification accuracy of 91% and 93.56% on the GTZAN and the ISMIR2004 Genre dataset, respectively. Both accuracies outperform any reported accuracy ever obtained by state of the art music genre classification algorithms in the aforementioned datasets.
  • Keywords
    audio recording; audio signal processing; compressed sensing; modulation; music; signal classification; signal representation; GTZAN; ISMIR2004 genre dataset; auditory temporal modulations; linear subspace dimensionality reduction techniques; music genre classification algorithms; music recordings; sparse representation-based classifiers; Accuracy; Dictionaries; Feature extraction; Modulation; Music; Principal component analysis; Training;
  • fLanguage
    English
  • Publisher
    ieee
  • Conference_Titel
    Signal Processing Conference, 2009 17th European
  • Conference_Location
    Glasgow
  • Print_ISBN
    978-161-7388-76-7
  • Type

    conf

  • Filename
    7077429