DocumentCode :
2740788
Title :
Using Projection Gradient Method to Train Linear Support Vector Machines
Author :
Niu, Lingfeng ; Shi, Yong
Author_Institution :
CAS Res. Center on Fictitious Econ. & Data Sci., Grad. Univ. of Chinese Acad. of Sci., Beijing, China
Volume :
3
fYear :
2010
fDate :
Aug. 31 2010-Sept. 3 2010
Firstpage :
207
Lastpage :
210
Abstract :
Linear Support Vector Machines(SVMs) have broad application in supervised classification problem with high dimensional feature space, such as text classification, word sense disambiguation, email spam detection and etc.. Considering the large volume of available training data, efficient training algorithm for linear SVMs draws many attention from the research community in recent years. Cutting-plane based method is one of the state-of-the-art training algorithms for linear SVMs. Within this cutting-plane framework, the quadratic programming(QP) subproblem, which consists of boundary constraints and a single inequality constraint, need to be solved at each iteration. This step is one of the most time consuming tasks in the whole method. In the current software, the QP subproblems are usually solved by the interior point method. In order to improve the efficiency of the cutting-plane based training algorithm, we transform the inequality constraint to an equation by introducing the slack variable and propose using projection gradient algorithm to solve the transformed QP subproblem. Compared with the existing method, the new algorithm has the following advantages. Firstly, because the special structure information in the subproblem is used carefully, the efficiency of solving the subproblem can be improved significantly. Secondly, through projecting the variables to the bound constraints explicitly, the variables that are not related to support vectors can be identified directly. Therefore, the rounding techniques, which is a necessary step in the widely used interior point method based solvers, is not required anymore. Experimental results on several public data sets also show the effectiveness and efficiency of our new algorithm.
Keywords :
gradient methods; pattern classification; quadratic programming; support vector machines; boundary constraint; cutting-plane based method; email spam detection; high dimensional feature space; interior point method; linear support vector machine training; projection gradient method; quadratic programming; single inequality constraint; slack variable; supervised classification problem; text classification; training algorithm; word sense disambiguation; Biological system modeling; Electronic mail; Machine learning; Support vector machines; Training; USA Councils; Vectors; Cutting Plane; Large Margin; Linear Support Vector Machines; Projection Gradient;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Web Intelligence and Intelligent Agent Technology (WI-IAT), 2010 IEEE/WIC/ACM International Conference on
Conference_Location :
Toronto, ON
Print_ISBN :
978-1-4244-8482-9
Electronic_ISBN :
978-0-7695-4191-4
Type :
conf
DOI :
10.1109/WI-IAT.2010.232
Filename :
5614639
Link To Document :
بازگشت