Title :
Comparison of MLP cost functions to dodge mislabeled training data
Author :
Nieminen, Paavo ; Karkkainen, Tommi
Author_Institution :
Dept. of Math. Inf. Technol., Univ. of Jyvaskyla, Jyvaskyla, Finland
Abstract :
Multilayer perceptrons (MLP) are often trained by minimizing the mean of squared errors (MSE), which is a sum of squared Euclidean norms of error vectors. Less common is to minimize the sum of Euclidean norms without squaring them. The latter approach, mean of non-squared errors (ME), bears implications from robust statistics. We carried out computational experiments to see if it would be notably better to train an MLP classifier by minimizing ME instead of MSE in the special case when training data contains class noise, i.e., when there is some mislabeling. Based on our experiments, we conclude that for small datasets containing class noise, ME could indeed be a very preferable choice, whereas for larger datasets it may not help.
Keywords :
mean square error methods; multilayer perceptrons; mean squared errors; mislabeled training data; multilayer perceptrons; squared Euclidean norms; Iris;
Conference_Titel :
Neural Networks (IJCNN), The 2010 International Joint Conference on
Conference_Location :
Barcelona
Print_ISBN :
978-1-4244-6916-1
DOI :
10.1109/IJCNN.2010.5596865