• DocumentCode
    2372276
  • Title

    Two new regularized AdaBoost algorithms

  • Author

    Yijun Sun ; Jian Li ; Hager, W.

  • Author_Institution
    Department of Electrical and Computer Engineering, University of Florida, Gainesville, FL 32611, USA
  • fYear
    2004
  • fDate
    16-18 Dec. 2004
  • Firstpage
    41
  • Lastpage
    48
  • Abstract
    AdaBoost rarely suffers from overfitting problems in low noise data cases. However, recent studies with highly noisy patterns clearly showed that overfitting can occur. A natural strategy to alleviate the problem is to penalize the distribution skewness in the learning process to prevent several hardest examples from spoiling decision boundaries. In this paper, we describe in detail how a penalty scheme can be pursued in the mathematical programming setting as well as in the Boosting setting. By using two smooth convex penalty functions, two new soft margin concepts are defined and two new regularized AdaBoost algorithms are proposed. The effectiveness of the proposed algorithms is demonstrated through a large scale experiment. Compared with other regularized AdaBoost algorithms, our methods can achieve at least the same or much better performances.
  • Keywords
    Algorithm design and analysis; Boosting; Large-scale systems; Linear approximation; Mathematical programming; Mathematics; Minimax techniques; Noise robustness; Performance evaluation; Sun;
  • fLanguage
    English
  • Publisher
    ieee
  • Conference_Titel
    Machine Learning and Applications, 2004. Proceedings. 2004 International Conference on
  • Conference_Location
    Louisville, Kentucky, USA
  • Print_ISBN
    0-7803-8823-2
  • Type

    conf

  • DOI
    10.1109/ICMLA.2004.1383492
  • Filename
    1383492