DocumentCode
303241
Title
Local minima and generalization
Author
Lawrence, Steve ; Tsoi, Ah Chung ; Giles, C. Lee
Author_Institution
NEC Res. Inst., Princeton, NJ, USA
Volume
1
fYear
1996
fDate
3-6 Jun 1996
Firstpage
371
Abstract
We consider a number of popular beliefs within the neural network community on the training and generalization behavior of multilayer perceptrons, and to some extent recurrent networks that: 1) the solution found is often close to the global minimum in terms of the magnitude of the error; 2) smaller networks generalize better than larger networks; and 3) the number of parameters in the network should be less than the number of data points in order to provide good generalization. For the tasks and methodology we consider, we show that: 1) the solution found is often significantly worse than the global minimum; 2) oversize networks can provide improved generalization due to their ability to find better solutions; and 3) the optimal number of parameters with respect to generalization error can be much larger than the number of data points
Keywords
backpropagation; generalisation (artificial intelligence); multilayer perceptrons; optimisation; recurrent neural nets; backpropagation; data points; generalization; global minima; multilayer perceptrons; recurrent networks; Australia; Backpropagation algorithms; Computer networks; Data engineering; Interference; Multi-layer neural network; Multilayer perceptrons; National electric code; Neural networks; Neurons;
fLanguage
English
Publisher
ieee
Conference_Titel
Neural Networks, 1996., IEEE International Conference on
Conference_Location
Washington, DC
Print_ISBN
0-7803-3210-5
Type
conf
DOI
10.1109/ICNN.1996.548920
Filename
548920
Link To Document