DocumentCode :
1724112
Title :
Norm-Induced Entropies for Decision Forests
Author :
Lassner, Christoph ; Lienhart, Rainer
fYear :
2015
Firstpage :
968
Lastpage :
975
Abstract :
The entropy measurement function is a central element of decision forest induction. The Shannon entropy and other generalized entropies such as the Renyi and Tsallis entropy are designed to fulfill the Khinchin-Shannon axioms. Whereas these axioms are appropriate for physical systems, they do not necessarily model well the artificial system of decision forest induction. In this paper, we show that when omitting two of the four axioms, every norm induces an entropy function. The remaining two axioms are sufficient to describe the requirements for an entropy function in the decision forest context. Furthermore, we introduce and analyze the p-norm-induced entropy, show relations to existing entropies and the relation to various heuristics that are commonly used for decision forest training. In experiments with classification, regression and the recently introduced Hough forests, we show how the discrete and differential form of the new entropy can be used for forest induction and how the functions can simply be fine tuned. The experiments indicate that the impact of the entropy function is limited, however can be a simple and useful post-processing step for optimizing decision forests for high performance applications.
Keywords :
decision trees; entropy; learning (artificial intelligence); Hough forests; classification; decision forest induction; decision forest training; entropy function; norm-induced entropies; p-norm-induced entropy; regression; Computer vision; Context; Entropy; Equations; Training; Vectors; Vegetation;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Applications of Computer Vision (WACV), 2015 IEEE Winter Conference on
Conference_Location :
Waikoloa, HI
Type :
conf
DOI :
10.1109/WACV.2015.134
Filename :
7045988
Link To Document :
بازگشت