Title :
Approximate l-Fold Cross-Validation with Least Squares SVM and Kernel Ridge Regression
Author :
Edwards, Richard E. ; Hao Zhang ; Parker, Lynne E. ; New, Joshua R.
Author_Institution :
Electr. Eng. & Comput. Sci., Univ. of Tennessee, Knoxville, TN, USA
Abstract :
Kernel methods have difficulties scaling to large modern data sets. The scalability issues are based on computational and memory requirements for working with a large matrix. These requirements have been addressed over the years by using low-rank kernel approximations or by improving the solvers´ scalability. However, Least Squares Support Vector Machines (LS-SVM), a popular SVM variant, and Kernel Ridge Regression still have several scalability issues. In particular, the O(n^3) computational complexity for solving a single model, and the overall computational complexity associated with tuning hyper parameters are still major problems. We address these problems by introducing an O(nlog n) approximate l-fold cross-validation method that uses a multi-level circulant matrix to approximate the kernel. In addition, we prove our algorithm´s computational complexity and present empirical runtimes on data sets with approximately one million data points. We also validate our approximate method´s effectiveness at selecting hyper parameters on real world and standard benchmark data sets. Lastly, we provide experimental results on using a multi level circulant kernel approximation to solve LS-SVM problems with hyper parameters selected using our method.
Keywords :
approximation theory; computational complexity; data handling; least squares approximations; matrix algebra; regression analysis; support vector machines; LS-SVM; approximate l-fold cross-validation; computational complexity; computational requirements; hyper parameter tuning; kernel methods; kernel ridge regression; least squares SVM; least squares support vector machines; low-rank kernel approximations; memory requirements; multilevel circulant kernel approximation; multilevel circulant matrix; real world benchmark data sets; scalability issues; standard benchmark data sets; Approximation algorithms; Computational complexity; Kernel; Least squares approximations; Runtime; Support vector machines;
Conference_Titel :
Machine Learning and Applications (ICMLA), 2013 12th International Conference on
Conference_Location :
Miami, FL
DOI :
10.1109/ICMLA.2013.18