DocumentCode :
1384965
Title :
On overfitting, generalization, and randomly expanded training sets
Author :
Karystinos, George N. ; Pados, Dimitris A.
Author_Institution :
Dept. of Electr. Eng., State Univ. of New York, Buffalo, NY, USA
Volume :
11
Issue :
5
fYear :
2000
fDate :
9/1/2000 12:00:00 AM
Firstpage :
1050
Lastpage :
1057
Abstract :
An algorithmic procedure is developed for the random expansion of a given training set to combat overfitting and improve the generalization ability of backpropagation trained multilayer perceptrons (MLPs). The training set is K-means clustered and locally most entropic colored Gaussian joint input-output probability density function estimates are formed per cluster. The number of clusters is chosen such that the resulting overall colored Gaussian mixture exhibits minimum differential entropy upon global cross-validated shaping. Numerical studies on real data and synthetic data examples drawn from the literature illustrate and support these theoretical developments
Keywords :
Gaussian distribution; backpropagation; generalisation (artificial intelligence); multilayer perceptrons; probability; K-means clustering; algorithmic procedure; backpropagation trained multilayer perceptrons; entropic colored Gaussian joint input-output probability density function estimates; global cross-validated shaping; minimum differential entropy; overfitting; random expansion; randomly expanded training sets; Backpropagation algorithms; Clustering algorithms; Clustering methods; Entropy; Minimization methods; Multilayer perceptrons; Neural networks; Probability density function; Stochastic processes; Testing;
fLanguage :
English
Journal_Title :
Neural Networks, IEEE Transactions on
Publisher :
ieee
ISSN :
1045-9227
Type :
jour
DOI :
10.1109/72.870038
Filename :
870038
Link To Document :
بازگشت