• DocumentCode
    303241
  • Title

    Local minima and generalization

  • Author

    Lawrence, Steve ; Tsoi, Ah Chung ; Giles, C. Lee

  • Author_Institution
    NEC Res. Inst., Princeton, NJ, USA
  • Volume
    1
  • fYear
    1996
  • fDate
    3-6 Jun 1996
  • Firstpage
    371
  • Abstract
    We consider a number of popular beliefs within the neural network community on the training and generalization behavior of multilayer perceptrons, and to some extent recurrent networks that: 1) the solution found is often close to the global minimum in terms of the magnitude of the error; 2) smaller networks generalize better than larger networks; and 3) the number of parameters in the network should be less than the number of data points in order to provide good generalization. For the tasks and methodology we consider, we show that: 1) the solution found is often significantly worse than the global minimum; 2) oversize networks can provide improved generalization due to their ability to find better solutions; and 3) the optimal number of parameters with respect to generalization error can be much larger than the number of data points
  • Keywords
    backpropagation; generalisation (artificial intelligence); multilayer perceptrons; optimisation; recurrent neural nets; backpropagation; data points; generalization; global minima; multilayer perceptrons; recurrent networks; Australia; Backpropagation algorithms; Computer networks; Data engineering; Interference; Multi-layer neural network; Multilayer perceptrons; National electric code; Neural networks; Neurons;
  • fLanguage
    English
  • Publisher
    ieee
  • Conference_Titel
    Neural Networks, 1996., IEEE International Conference on
  • Conference_Location
    Washington, DC
  • Print_ISBN
    0-7803-3210-5
  • Type

    conf

  • DOI
    10.1109/ICNN.1996.548920
  • Filename
    548920