DocumentCode
396741
Title
Model selection for k-nearest neighbors regression using VC bounds
Author
Cherkassky, Vladimir ; Ma, Yunqian ; Tang, Jun
Author_Institution
Dept. of Electr. & Comput. Eng., Minnesota Univ., USA
Volume
2
fYear
2003
fDate
20-24 July 2003
Firstpage
1143
Abstract
We discuss an analytic model selection for k-nearest neighbors regression method using VC generalization bounds. Whereas existing implementations of k-nn regression estimate the model complexity as n/k, where n is the number of samples, we propose a new model complexity estimate. The proposed new complexity index used as the VC-dimension in VC bounds yields a new analytic method for model selection. Empirical results for low dimensional and high dimensional data sets indicate that the proposed model selection approach provides accurate model selection that is consistently better than the previously used complexity measure. In fact, prediction accuracy of the proposed analytic method is similar to resampling (cross-validation) approach for optimal selection of k.
Keywords
learning (artificial intelligence); regression analysis; VC generalization bounds; analytic method; complexity index; cross-validation approach; k-nearest neighbors regression; k-nn regression; model selection approach; optimal selection; prediction accuracy; resampling approach; Accuracy; Learning systems; Loss measurement; Multidimensional systems; Nearest neighbor searches; Parameter estimation; Predictive models; Risk analysis; Training data; Virtual colonoscopy;
fLanguage
English
Publisher
ieee
Conference_Titel
Neural Networks, 2003. Proceedings of the International Joint Conference on
ISSN
1098-7576
Print_ISBN
0-7803-7898-9
Type
conf
DOI
10.1109/IJCNN.2003.1223852
Filename
1223852
Link To Document