Title :
Effect analysis of dimension reduction on support vector machines
Author :
Zhu, Muhua ; Zhu, Jingbo ; Chen, Wenliang
Author_Institution :
Natural Language Process. Lab., Northeastern Univ., Shenyang, China
fDate :
30 Oct.-1 Nov. 2005
Abstract :
Support vector machines (SVMs) are extensively used for text categorization, and dimension reduction is optional, not imperative for SVMs. But in some time-critical applications, dimension reduction of feature space is still necessary. In this paper, universal dimension reduction methods: feature selection and feature extraction, are applied to SVMs. At the same time, we also examine the influence of different kernel functions on the performance of dimension reduction. In the feature selection case, experimental results show that when the linear kernel is used for SVMs, the performance is close to the baseline system, sometimes even better, and when nonlinear kernel is employed, feature selection methods get the performance decrease sharply. On the contrary, principal component analysis (PCA), one of feature extraction methods, gets excellent performance with both linear and nonlinear kernel functions.
Keywords :
classification; feature extraction; principal component analysis; support vector machines; text analysis; dimension reduction; feature extraction; feature selection; kernel functions; principal component analysis; support vector machines; text categorization; Feature extraction; Indexing; Kernel; Polynomials; Principal component analysis; Risk management; Supervised learning; Support vector machines; Text categorization; Time factors;
Conference_Titel :
Natural Language Processing and Knowledge Engineering, 2005. IEEE NLP-KE '05. Proceedings of 2005 IEEE International Conference on
Print_ISBN :
0-7803-9361-9
DOI :
10.1109/NLPKE.2005.1598806