DocumentCode :
288333
Title :
An incremental learning algorithm that optimizes network size and sample size in one trial
Author :
Zhang, Byoung-Tak
Author_Institution :
German Nat. Res. Center for Comput. Sci., St. Augustin, Germany
Volume :
1
fYear :
1994
fDate :
27 Jun-2 Jul 1994
Firstpage :
215
Abstract :
A constructive learning algorithm is described that builds a feedforward neural network with an optimal number of hidden units to balance convergence and generalization. The method starts with a small training set and a small network, and expands the training set incrementally after training. If the training does not converge, the network grows incrementally to increase its learning capacity. This process, called selective learning with flexible neural architectures (SELF), results in a construction of an optimal size network for learning all the given data using only a minimal subset of them. The author shows that the network size optimization combined with active example selection generalizes significantly better and converges faster than conventional methods
Keywords :
convergence; feedforward neural nets; generalisation (artificial intelligence); learning (artificial intelligence); active example selection; convergence; feedforward neural network; generalization; incremental learning algorithm; learning capacity; network size optimization; Buildings; Computer science; Convergence; Electronic mail; Feedforward neural networks; Intelligent networks; Multilayer perceptrons; Neural networks; Optimization methods; Radial basis function networks;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Neural Networks, 1994. IEEE World Congress on Computational Intelligence., 1994 IEEE International Conference on
Conference_Location :
Orlando, FL
Print_ISBN :
0-7803-1901-X
Type :
conf
DOI :
10.1109/ICNN.1994.374165
Filename :
374165
Link To Document :
بازگشت