DocumentCode :
1465928
Title :
Structural Regularized Support Vector Machine: A Framework for Structural Large Margin Classifier
Author :
Xue, Hui ; Chen, Songcan ; Yang, Qiang
Author_Institution :
Sch. of Comput. Sci. & Eng., Southeast Univ., Nanjing, China
Volume :
22
Issue :
4
fYear :
2011
fDate :
4/1/2011 12:00:00 AM
Firstpage :
573
Lastpage :
587
Abstract :
Support vector machine (SVM), as one of the most popular classifiers, aims to find a hyperplane that can separate two classes of data with maximal margin. SVM classifiers are focused on achieving more separation between classes than exploiting the structures in the training data within classes. However, the structural information, as an implicit prior knowledge, has recently been found to be vital for designing a good classifier in different real-world problems. Accordingly, using as much prior structural information in data as possible to help improve the generalization ability of a classifier has yielded a class of effective structural large margin classifiers, such as the structured large margin machine (SLMM) and the Laplacian support vector machine (LapSVM). In this paper, we unify these classifiers into a common framework from the concept of “structural granularity” and the formulation for optimization problems. We exploit the quadratic programming (QP) and second-order cone programming (SOCP) methods, and derive a novel large margin classifier, we call the new classifier the structural regularized support vector machine (SRSVM). Unlike both SLMM at the cross of the cluster granularity and SOCP and LapSVM at the cross of the point granularity and QP, SRSVM is located at the cross of the cluster granularity and QP and thus follows the same optimization formulation as LapSVM to overcome large computational complexity and non-sparse solution in SLMM. In addition, it integrates the compactness within classes with the separability between classes simultaneously. Furthermore, it is possible to derive generalization bounds for these algorithms by using eigenvalue analysis of the kernel matrices. Experimental results demonstrate that SRSVM is often superior in classification and generalization performances to the state-of-the-art algorithms in the framework, both with the same and different structural granularities.
Keywords :
computational complexity; eigenvalues and eigenfunctions; learning (artificial intelligence); matrix algebra; pattern classification; quadratic programming; support vector machines; Laplacian support vector machine; SVM classifiers; cluster granularity; computational complexity; eigenvalue analysis; kernel matrices; machine learning; maximal margin; optimization problems; quadratic programming method; second-order cone programming method; structural granularity; structural information; structural large margin classifier; structural regularized support vector machine; Clustering algorithms; Covariance matrix; Ellipsoids; Kernel; Manifolds; Optimization; Support vector machines; Generalization bound; machine learning; structural granularity; support vector machine; Algorithms; Artificial Intelligence; Classification; Cluster Analysis; Databases, Factual; Humans; Neural Networks (Computer); Pattern Recognition, Automated;
fLanguage :
English
Journal_Title :
Neural Networks, IEEE Transactions on
Publisher :
ieee
ISSN :
1045-9227
Type :
jour
DOI :
10.1109/TNN.2011.2108315
Filename :
5724308
Link To Document :
بازگشت