Title :
Generalized neural trees for pattern classification
Author :
Foresti, Gian Luca ; Micheloni, Christian
Author_Institution :
Dept. of Math. & Comput. Sci., Udine Univ., Italy
fDate :
11/1/2002 12:00:00 AM
Abstract :
In this paper, a new neural tree (NT) model, the generalized NT (GNT), is presented. The main novelty of the GNT consists in the definition of a new training rule that performs an overall optimization of the tree. Each time the tree is increased by a new level, the whole tree is reevaluated. The training rule uses a weight correction strategy that takes into account the entire tree structure, and it applies a normalization procedure to the activation values of each node such that these values can be interpreted as a probability. The weight connection updating is calculated by minimizing a cost function, which represents a measure of the overall probability of correct classification. Significant results on both synthetic and real data have been obtained by comparing the classification performances among multilayer perceptrons (MLPs), NTs, and GNTs. In particular, the GNT model displays good classification performances for training sets having complex distributions. Moreover, its particular structure provides an easily probabilistic interpretation of the pattern classification task and allows growing small neural trees with good generalization properties.
Keywords :
multilayer perceptrons; neural nets; pattern classification; search problems; generalized neural trees; multilayer perceptrons; neural tree model; normalization procedure; pattern classification; search methods; training rule; tree structure; weight connection updating; weight correction strategy; Classification tree analysis; Cost function; Displays; Multilayer perceptrons; Neural networks; Neurons; Pattern classification; Performance evaluation; Probability; Tree data structures;
Journal_Title :
Neural Networks, IEEE Transactions on
DOI :
10.1109/TNN.2002.804290