DocumentCode :
2496591
Title :
Comparison of MLP cost functions to dodge mislabeled training data
Author :
Nieminen, Paavo ; Karkkainen, Tommi
Author_Institution :
Dept. of Math. Inf. Technol., Univ. of Jyvaskyla, Jyvaskyla, Finland
fYear :
2010
fDate :
18-23 July 2010
Firstpage :
1
Lastpage :
7
Abstract :
Multilayer perceptrons (MLP) are often trained by minimizing the mean of squared errors (MSE), which is a sum of squared Euclidean norms of error vectors. Less common is to minimize the sum of Euclidean norms without squaring them. The latter approach, mean of non-squared errors (ME), bears implications from robust statistics. We carried out computational experiments to see if it would be notably better to train an MLP classifier by minimizing ME instead of MSE in the special case when training data contains class noise, i.e., when there is some mislabeling. Based on our experiments, we conclude that for small datasets containing class noise, ME could indeed be a very preferable choice, whereas for larger datasets it may not help.
Keywords :
mean square error methods; multilayer perceptrons; mean squared errors; mislabeled training data; multilayer perceptrons; squared Euclidean norms; Iris;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Neural Networks (IJCNN), The 2010 International Joint Conference on
Conference_Location :
Barcelona
ISSN :
1098-7576
Print_ISBN :
978-1-4244-6916-1
Type :
conf
DOI :
10.1109/IJCNN.2010.5596865
Filename :
5596865
Link To Document :
بازگشت