DocumentCode :
1408819
Title :
Optimal partitioning for classification and regression trees
Author :
Chou, Philip A.
Author_Institution :
Dept. of Electr. Eng., Standford Univ., CA, USA
Volume :
13
Issue :
4
fYear :
1991
fDate :
4/1/1991 12:00:00 AM
Firstpage :
340
Lastpage :
354
Abstract :
An iterative algorithm that finds a locally optimal partition for an arbitrary loss function, in time linear in N for each iteration is presented. The algorithm is a K-means-like clustering algorithm that uses as its distance measure a generalization of Kullback´s information divergence. Moreover, it is proven that the globally optimal partition must satisfy a nearest neighbour condition using divergence as the distance measure. These results generalize similar results of L. Breiman et al. (1984) to an arbitrary number of classes or regression variables and to an arbitrary number of bills. Experimental results on a text-to-speech example are provided and additional applications of the algorithm, including the design of variable combinations, surrogate splits, composite nodes, and decision graphs, are suggested
Keywords :
decision theory; iterative methods; speech recognition; trees (mathematics); Kullback´s information divergence; clustering algorithm; composite nodes; decision graphs; iterative algorithm; partitioning; regression trees; speech recognition; surrogate splits; text-to-speech; Bars; Classification tree analysis; Clustering algorithms; Decision trees; Iterative algorithms; Nearest neighbor searches; Optical character recognition software; Partitioning algorithms; Regression tree analysis; Testing;
fLanguage :
English
Journal_Title :
Pattern Analysis and Machine Intelligence, IEEE Transactions on
Publisher :
ieee
ISSN :
0162-8828
Type :
jour
DOI :
10.1109/34.88569
Filename :
88569
Link To Document :
بازگشت