Title :
Characterizing the error function of a neural network
Author :
Moore, Barbara ; Fogaca, Marcelo ; Kramer, Alan
Author_Institution :
MIT AI Lab., Cambridge, MA, USA
Abstract :
Several means are discussed for exploring the error function of a multilayer, feedforward neural network. In particular, hyperplane configurations over time and the generalization of the network function to a region of the input space are considered. The analysis provides explanations for empirically observed phenomena, such as flat spots in the energy surface, and improvements in performance with extra hidden units. The implementation of the back-propagation network training algorithm on the massively parallel connection machine computer is discussed. Results are presented of a scaling experiment in which the effect on learning time of the number of hidden units is studied. Also, results are presented of an experiment on adjusting the length of the initial random weight vectors. The explorations of the shape of the energy function in weight space and the hyperplane configuration are discussed. Several methods for obtaining speedup in training times based on the analyses are suggested
Keywords :
neural nets; parallel processing; back-propagation network training algorithm; error function; flat spots; hyperplane configurations; massively parallel connection machine computer; neural network; random weight vectors; scaling experiment; Algorithm design and analysis; Artificial intelligence; Computer errors; Computer networks; Concurrent computing; Feedforward neural networks; Multi-layer neural network; Neural networks; Shape; Transfer functions;
Conference_Titel :
Frontiers of Massively Parallel Computation, 1988. Proceedings., 2nd Symposium on the Frontiers of
Conference_Location :
Fairfax, VA
Print_ISBN :
0-8186-5892-4
DOI :
10.1109/FMPC.1988.47412