DocumentCode
2209239
Title
Monotone Relabeling in Ordinal Classification
Author
Feelders, Ad
Author_Institution
Univ. Utrecht, Utrecht, Netherlands
fYear
2010
fDate
13-17 Dec. 2010
Firstpage
803
Lastpage
808
Abstract
In many applications of data mining we know beforehand that the response variable should be increasing (or decreasing) in the attributes. Such relations between response and attributes are called monotone. In this paper we present a new algorithm to compute an optimal monotone classification of a data set for convex loss functions. Moreover, we show how the algorithm can be extended to compute all optimal monotone classifications with little additional effort. Monotone relabeling is useful for at least two reasons. Firstly, models trained on relabeled data sets often have better predictive performance than models trained on the original data. Secondly, relabeling is an important building block for the construction of monotone classifiers. We apply the new algorithm to investigate the effect on the prediction error of relabeling the training sample for k nearest neighbour classification and classification trees. In contrast to previous work in this area, we consider all optimal monotone relabelings. The results show that, for small training samples, relabeling the training data results in significantly better predictive performance.
Keywords
data mining; learning (artificial intelligence); pattern classification; classification trees; convex loss functions; data mining; k nearest neighbour classification; monotone classification; monotone relabeling; ordinal classification;
fLanguage
English
Publisher
ieee
Conference_Titel
Data Mining (ICDM), 2010 IEEE 10th International Conference on
Conference_Location
Sydney, NSW
ISSN
1550-4786
Print_ISBN
978-1-4244-9131-5
Electronic_ISBN
1550-4786
Type
conf
DOI
10.1109/ICDM.2010.92
Filename
5694042
Link To Document