• DocumentCode
    2833640
  • Title

    Random decision forests

  • Author

    Ho, Tin Kam

  • Author_Institution
    AT&T Bell Labs., Murray Hill, NJ, USA
  • Volume
    1
  • fYear
    1995
  • fDate
    14-16 Aug 1995
  • Firstpage
    278
  • Abstract
    Decision trees are attractive classifiers due to their high execution speed. But trees derived with traditional methods often cannot be grown to arbitrary complexity for possible loss of generalization accuracy on unseen data. The limitation on complexity usually means suboptimal accuracy on training data. Following the principles of stochastic modeling, we propose a method to construct tree-based classifiers whose capacity can be arbitrarily expanded for increases in accuracy for both training and unseen data. The essence of the method is to build multiple trees in randomly selected subspaces of the feature space. Trees in, different subspaces generalize their classification in complementary ways, and their combined classification can be monotonically improved. The validity of the method is demonstrated through experiments on the recognition of handwritten digits
  • Keywords
    decision theory; handwriting recognition; optical character recognition; complexity; decision trees; generalization accuracy; handwritten digits; random decision forests; stochastic modeling; suboptimal accuracy; tree-based classifiers; Classification tree analysis; Decision trees; Handwriting recognition; Hidden Markov models; Multilayer perceptrons; Optimization methods; Stochastic processes; Testing; Tin; Training data;
  • fLanguage
    English
  • Publisher
    ieee
  • Conference_Titel
    Document Analysis and Recognition, 1995., Proceedings of the Third International Conference on
  • Conference_Location
    Montreal, Que.
  • Print_ISBN
    0-8186-7128-9
  • Type

    conf

  • DOI
    10.1109/ICDAR.1995.598994
  • Filename
    598994