DocumentCode
2372276
Title
Two new regularized AdaBoost algorithms
Author
Yijun Sun ; Jian Li ; Hager, W.
Author_Institution
Department of Electrical and Computer Engineering, University of Florida, Gainesville, FL 32611, USA
fYear
2004
fDate
16-18 Dec. 2004
Firstpage
41
Lastpage
48
Abstract
AdaBoost rarely suffers from overfitting problems in low noise data cases. However, recent studies with highly noisy patterns clearly showed that overfitting can occur. A natural strategy to alleviate the problem is to penalize the distribution skewness in the learning process to prevent several hardest examples from spoiling decision boundaries. In this paper, we describe in detail how a penalty scheme can be pursued in the mathematical programming setting as well as in the Boosting setting. By using two smooth convex penalty functions, two new soft margin concepts are defined and two new regularized AdaBoost algorithms are proposed. The effectiveness of the proposed algorithms is demonstrated through a large scale experiment. Compared with other regularized AdaBoost algorithms, our methods can achieve at least the same or much better performances.
Keywords
Algorithm design and analysis; Boosting; Large-scale systems; Linear approximation; Mathematical programming; Mathematics; Minimax techniques; Noise robustness; Performance evaluation; Sun;
fLanguage
English
Publisher
ieee
Conference_Titel
Machine Learning and Applications, 2004. Proceedings. 2004 International Conference on
Conference_Location
Louisville, Kentucky, USA
Print_ISBN
0-7803-8823-2
Type
conf
DOI
10.1109/ICMLA.2004.1383492
Filename
1383492
Link To Document