DocumentCode
1942544
Title
Optimizing 0/1 Loss for Perceptrons by Random Coordinate Descent
Author
Li, Ling ; Lin, Hsuan-Tien
Author_Institution
California Inst. of Technol., Pasadena
fYear
2007
fDate
12-17 Aug. 2007
Firstpage
749
Lastpage
754
Abstract
The 0/1 loss is an important cost function for perceptrons. Nevertheless it cannot be easily minimized by most existing perceptron learning algorithms. In this paper, we propose a family of random coordinate descent algorithms to directly minimize the 0/1 loss for perceptrons, and prove their convergence. Our algorithms are computationally efficient, and usually achieve the lowest 0/1 loss compared with other algorithms. Such advantages make them favorable for nonseparable real-world problems. Experiments show that our algorithms are especially useful for ensemble learning, and could achieve the lowest test error for many complex data sets when coupled with AdaBoost.
Keywords
perceptrons; random processes; perceptron learning algorithm; random coordinate descent; Algorithm design and analysis; Biological neural networks; Brain modeling; Convergence; Cost function; Iterative algorithms; Learning systems; Minimization methods; Support vector machines; Testing;
fLanguage
English
Publisher
ieee
Conference_Titel
Neural Networks, 2007. IJCNN 2007. International Joint Conference on
Conference_Location
Orlando, FL
ISSN
1098-7576
Print_ISBN
978-1-4244-1379-9
Electronic_ISBN
1098-7576
Type
conf
DOI
10.1109/IJCNN.2007.4371051
Filename
4371051
Link To Document