Title :
A hybrid optimization method for acceleration of building linear classification models
Author :
Junchao Lv ; Qiang Wang ; Huang, Joshua Zhexue
Author_Institution :
Shenzhen Key Lab. of High Performance Data Min., Shenzhen Univ. Town, Shenzhen, China
Abstract :
Linear classification is an important technique in machine learning and data mining, and development of fast optimization methods for training linear classification models is a hot research topic. Stochastic gradient descent (SGD) can achieve relatively good results quickly, but unstable to converge. Limited-memory BFGS (L-BFGS) method converges, but takes a long time to train the model, as it needs to compute the gradient from the entire data set to make an update. In this paper, we investigate a hybrid method that integrates SGD and L-BFGS into a new optimization process SGD-LBFGS to take advantages of both optimization methods. In SGD-LBFGS, SGD is used to run initial iterations to obtain a suboptimal result, and then L-BFGS takes over to continue the optimization process until the process converges and a better model is built. We present a theoretical result to prove that SGD-LBFGS converges faster than SGD and L-BFGS. Experiment analysis on 6 real world data sets have shown that SGD-LBFGS converged 77% faster than L-BFGS on average and demonstrated more stable results than SGD.
Keywords :
data mining; gradient methods; iterative methods; learning (artificial intelligence); optimisation; pattern classification; stochastic processes; building linear classification model acceleration; data mining; fast optimization methods; hybrid optimization method; limited-memory BFGS method; machine learning; optimization process SGD-LBFGS; stochastic gradient descent; Convergence; Learning systems; Linear programming; Logistics; Optimization methods; Training;
Conference_Titel :
Neural Networks (IJCNN), The 2013 International Joint Conference on
Conference_Location :
Dallas, TX
Print_ISBN :
978-1-4673-6128-6
DOI :
10.1109/IJCNN.2013.6707017