Title :
Two new regularized AdaBoost algorithms
Author :
Yijun Sun ; Jian Li ; Hager, W.
Author_Institution :
Department of Electrical and Computer Engineering, University of Florida, Gainesville, FL 32611, USA
Abstract :
AdaBoost rarely suffers from overfitting problems in low noise data cases. However, recent studies with highly noisy patterns clearly showed that overfitting can occur. A natural strategy to alleviate the problem is to penalize the distribution skewness in the learning process to prevent several hardest examples from spoiling decision boundaries. In this paper, we describe in detail how a penalty scheme can be pursued in the mathematical programming setting as well as in the Boosting setting. By using two smooth convex penalty functions, two new soft margin concepts are defined and two new regularized AdaBoost algorithms are proposed. The effectiveness of the proposed algorithms is demonstrated through a large scale experiment. Compared with other regularized AdaBoost algorithms, our methods can achieve at least the same or much better performances.
Keywords :
Algorithm design and analysis; Boosting; Large-scale systems; Linear approximation; Mathematical programming; Mathematics; Minimax techniques; Noise robustness; Performance evaluation; Sun;
Conference_Titel :
Machine Learning and Applications, 2004. Proceedings. 2004 International Conference on
Conference_Location :
Louisville, Kentucky, USA
Print_ISBN :
0-7803-8823-2
DOI :
10.1109/ICMLA.2004.1383492