DocumentCode :
2933080
Title :
Decision Tree Ensemble: Small Heterogeneous Is Better Than Large Homogeneous
Author :
Gashler, Mike ; Giraud-Carrier, Christophe ; Martinez, Tony
Author_Institution :
Department of Computer Science, Brigham Young University, Provo, UT, U.S.A.
fYear :
2008
fDate :
11-13 Dec. 2008
Firstpage :
900
Lastpage :
905
Abstract :
Using decision trees that split on randomly selected attributes is one way to increase the diversity within an ensemble of decision trees. Another approach increases diversity by combining multiple tree algorithms. The random forest approach has become popular because it is simple and yields good results with common datasets. We present a technique that combines heterogeneous tree algorithms and contrast it with homogeneous forest algorithms. Our results indicate that random forests do poorly when faced with irrelevant attributes, while our heterogeneous technique handles them robustly. Further, we show that large ensembles of random trees are more susceptible to diminishing returns than our technique. We are able to obtain better results across a large number of common datasets with a significantly smaller ensemble.
Keywords :
Accuracy; Algorithm design and analysis; Application software; Bagging; Computer science; Decision trees; Diversity reception; Machine learning; Robustness; Training data;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Machine Learning and Applications, 2008. ICMLA '08. Seventh International Conference on
Conference_Location :
San Diego, CA
Print_ISBN :
978-0-7695-3495-4
Type :
conf
DOI :
10.1109/ICMLA.2008.154
Filename :
4796917
Link To Document :
بازگشت