DocumentCode :
401684
Title :
Using mutual information for selecting continuous-valued attribute in decision tree learning
Author :
Li, Hua ; Wang, Xi-zhao ; Li, Yong
Author_Institution :
Fac. of Math. & Comput. Sci., Hebei Univ., China
Volume :
3
fYear :
2003
fDate :
2-5 Nov. 2003
Firstpage :
1496
Abstract :
In this paper, we proposed a learning algorithm using the information entropy minimization heuristic and mutual information entropy heuristic to select expanded attributes. For a data set of which the values of condition attributes are continuous, most of the current decision trees learning algorithms often select the previously selected attributes for branching. The repeated selection limits the accuracy of training and testing and the structure of decision trees may become complex. So in the selection of attributes, the previously selected attributes and the other attributes, which have high correlation to the previously selected attributes, should not be selected again. Here, we use mutual information to avoid selecting the previously selected attributes in the generation of decision trees and our test results show that this method can obtain good performance.
Keywords :
decision trees; entropy; heuristic programming; learning (artificial intelligence); minimisation; continuous-valued attribute; decision tree learning; information entropy minimization heuristic; mutual information entropy heuristic; Computer science; Decision trees; Information entropy; Information theory; Machine learning; Machine learning algorithms; Mathematics; Mutual information; Stochastic processes; Testing;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Machine Learning and Cybernetics, 2003 International Conference on
Print_ISBN :
0-7803-8131-9
Type :
conf
DOI :
10.1109/ICMLC.2003.1259731
Filename :
1259731
Link To Document :
بازگشت