Title :
An adaptive back-propagation learning method: A preliminary study for incremental neural networks
Author :
Chen, Hown-Wen ; Soo, Von-Wun
Author_Institution :
Inst. of Comput. Sci., Nat. Tsing Hua Univ., Hsinchu, Taiwan
Abstract :
The authors apply the concept of minimizing weight sensitivity cost and training square-error functions using gradient descent optimization techniques, and they obtain a novel supervised backpropagation learning algorithm on a biased two-layered perceptron. In addition to illustrating the conflict locality of an inserted training instance with respect to previous training data, they point out that this adaptive learning method can get a network with a measurable generalization ability. This work can also be extended to an incremental network in which no training instances are needed to be remembered
Keywords :
backpropagation; feedforward neural nets; learning (artificial intelligence); adaptive backpropagation learning method; biased two-layered perceptron; conflict locality; gradient descent optimization; incremental neural networks; measurable generalization; minimizing weight sensitivity cost; training square-error functions; Application software; Artificial neural networks; Computer science; Cost function; Humans; Learning systems; Load forecasting; Multilayer perceptrons; Neural networks; Training data;
Conference_Titel :
Neural Networks, 1992. IJCNN., International Joint Conference on
Conference_Location :
Baltimore, MD
Print_ISBN :
0-7803-0559-0
DOI :
10.1109/IJCNN.1992.287103