DocumentCode :
907845
Title :
A support vector machine with a hybrid kernel and minimal Vapnik-Chervonenkis dimension
Author :
Tan, Ying ; Wang, Jun
Author_Institution :
Dept. of Electron. Eng. & Inf. Sci., Univ. of Sci. & Technol. of China, Hefei, China
Volume :
16
Issue :
4
fYear :
2004
fDate :
4/1/2004 12:00:00 AM
Firstpage :
385
Lastpage :
395
Abstract :
We present a mechanism to train support vector machines (SVMs) with a hybrid kernel and minimal Vapnik-Chervonenkis (VC) dimension. After describing the VC dimension of sets of separating hyperplanes in a high-dimensional feature space produced by a mapping related to kernels from the input space, we proposed an optimization criterion to design SVMs by minimizing the upper bound of the VC dimension. This method realizes a structural risk minimization and utilizes a flexible kernel function such that a superior generalization over test data can be obtained. In order to obtain a flexible kernel function, we develop a hybrid kernel function and a sufficient condition to be an admissible Mercer kernel based on common Mercer kernels (polynomial, radial basis function, two-layer neural network, etc.). The nonnegative combination coefficients and parameters of the hybrid kernel are determined subject to the minimal upper bound of the VC dimension of the learning machine. The use of the hybrid kernel results in a better performance than those with a single common kernel. Experimental results are discussed to illustrate the proposed method and show that the SVM with the hybrid kernel outperforms that with a single common kernel in terms of generalization power.
Keywords :
generalisation (artificial intelligence); learning (artificial intelligence); minimisation; pattern recognition; radial basis function networks; support vector machines; SVM design; VC dimension; admissible Mercer kernel; generalization power; hybrid kernel function; hyperplane separation; learning machine; minimal Vapnik-Chervonenkis dimension; neural network; nonnegative combination coefficients; optimization criterion; radial basis function; structural risk minimization; support vector machine; test data; Design optimization; Kernel; Neural networks; Polynomials; Risk management; Sufficient conditions; Support vector machines; Testing; Upper bound; Virtual colonoscopy;
fLanguage :
English
Journal_Title :
Knowledge and Data Engineering, IEEE Transactions on
Publisher :
ieee
ISSN :
1041-4347
Type :
jour
DOI :
10.1109/TKDE.2004.1269664
Filename :
1269664
Link To Document :
بازگشت