• DocumentCode
    2489401
  • Title

    Feature selection for support vector regression via Kernel penalization

  • Author

    Maldonado, Sebastián ; Weber, Richard

  • Author_Institution
    Dept. of Ind. Eng., Univ. of Chile, Santiago, Chile
  • fYear
    2010
  • fDate
    18-23 July 2010
  • Firstpage
    1
  • Lastpage
    7
  • Abstract
    This paper presents a novel feature selection approach (KP-SVR) that determines a non-linear regression function with minimal error and simultaneously minimizes the number of features by penalizing their use in the dual formulation of SVR. The approach optimizes the width of an anisotropic RBF Kernel using an iterative algorithm based on the gradient descent method, eliminating features that have low relevance for the regression model. Our approach presents an explicit stopping criterion, indicating clearly when eliminating further features begins to affect negatively the model´s performance. Experiments with two real-world benchmark problems demonstrate that our approach accomplishes the best performance compared to well-known feature selection methods using consistently a small number of features.
  • Keywords
    gradient methods; regression analysis; support vector machines; anisotropic RBF Kernel; explicit stopping criterion; feature selection methods; gradient descent method; iterative algorithm; kernel penalization; nonlinear regression function; support vector machines; support vector regression; Feature extraction; Kernel; Optimization; Polynomials; Predictive models; Support vector machines; Training;
  • fLanguage
    English
  • Publisher
    ieee
  • Conference_Titel
    Neural Networks (IJCNN), The 2010 International Joint Conference on
  • Conference_Location
    Barcelona
  • ISSN
    1098-7576
  • Print_ISBN
    978-1-4244-6916-1
  • Type

    conf

  • DOI
    10.1109/IJCNN.2010.5596488
  • Filename
    5596488