• DocumentCode
    1942544
  • Title

    Optimizing 0/1 Loss for Perceptrons by Random Coordinate Descent

  • Author

    Li, Ling ; Lin, Hsuan-Tien

  • Author_Institution
    California Inst. of Technol., Pasadena
  • fYear
    2007
  • fDate
    12-17 Aug. 2007
  • Firstpage
    749
  • Lastpage
    754
  • Abstract
    The 0/1 loss is an important cost function for perceptrons. Nevertheless it cannot be easily minimized by most existing perceptron learning algorithms. In this paper, we propose a family of random coordinate descent algorithms to directly minimize the 0/1 loss for perceptrons, and prove their convergence. Our algorithms are computationally efficient, and usually achieve the lowest 0/1 loss compared with other algorithms. Such advantages make them favorable for nonseparable real-world problems. Experiments show that our algorithms are especially useful for ensemble learning, and could achieve the lowest test error for many complex data sets when coupled with AdaBoost.
  • Keywords
    perceptrons; random processes; perceptron learning algorithm; random coordinate descent; Algorithm design and analysis; Biological neural networks; Brain modeling; Convergence; Cost function; Iterative algorithms; Learning systems; Minimization methods; Support vector machines; Testing;
  • fLanguage
    English
  • Publisher
    ieee
  • Conference_Titel
    Neural Networks, 2007. IJCNN 2007. International Joint Conference on
  • Conference_Location
    Orlando, FL
  • ISSN
    1098-7576
  • Print_ISBN
    978-1-4244-1379-9
  • Electronic_ISBN
    1098-7576
  • Type

    conf

  • DOI
    10.1109/IJCNN.2007.4371051
  • Filename
    4371051