DocumentCode :
303241
Title :
Local minima and generalization
Author :
Lawrence, Steve ; Tsoi, Ah Chung ; Giles, C. Lee
Author_Institution :
NEC Res. Inst., Princeton, NJ, USA
Volume :
1
fYear :
1996
fDate :
3-6 Jun 1996
Firstpage :
371
Abstract :
We consider a number of popular beliefs within the neural network community on the training and generalization behavior of multilayer perceptrons, and to some extent recurrent networks that: 1) the solution found is often close to the global minimum in terms of the magnitude of the error; 2) smaller networks generalize better than larger networks; and 3) the number of parameters in the network should be less than the number of data points in order to provide good generalization. For the tasks and methodology we consider, we show that: 1) the solution found is often significantly worse than the global minimum; 2) oversize networks can provide improved generalization due to their ability to find better solutions; and 3) the optimal number of parameters with respect to generalization error can be much larger than the number of data points
Keywords :
backpropagation; generalisation (artificial intelligence); multilayer perceptrons; optimisation; recurrent neural nets; backpropagation; data points; generalization; global minima; multilayer perceptrons; recurrent networks; Australia; Backpropagation algorithms; Computer networks; Data engineering; Interference; Multi-layer neural network; Multilayer perceptrons; National electric code; Neural networks; Neurons;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Neural Networks, 1996., IEEE International Conference on
Conference_Location :
Washington, DC
Print_ISBN :
0-7803-3210-5
Type :
conf
DOI :
10.1109/ICNN.1996.548920
Filename :
548920
Link To Document :
بازگشت