DocumentCode :
3565867
Title :
Training self-configuring backpropagation networks
Author :
Bryant, Garnett W.
Author_Institution :
Harry Diamond Labs., Adelphi, MD, USA
Volume :
1
fYear :
1992
Firstpage :
365
Abstract :
A generalization of the generalized delta rule (GDR) for error backpropagation in feedforward networks is presented. The generalization applies to networks with any topology of connections between nodes and an activation function at each node that is any differentiable function of its inputs and of the parameters to be adjusted by error backpropagation. The generalized GDR is used to implement networks in which the strength (magnitude of the activation function) of each node is adjusted during training. These networks are trained by minimizing the training error plus cost functions for the node strengths. Self-configuring networks are implemented by use of cost functions which have minimum cost when the node is off (zero magnitude) or on (unit magnitude). Simulations show that a self-configuring network with all nodes initially inactive turns on nodes one by one during training. Simulations show that a self-configuring network can also deactivate nodes during training. Methods for training self-configuring networks are discussed
Keywords :
backpropagation; feedforward neural nets; knowledge based systems; activation function; cost functions; feedforward networks; generalized delta rule; self-configuring backpropagation networks training; training error; Backpropagation; Cost function; Feedforward neural networks; Feeds; Laboratories; Network topology; Neural networks; Physics; Training data;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Neural Networks, 1992. IJCNN., International Joint Conference on
Print_ISBN :
0-7803-0559-0
Type :
conf
DOI :
10.1109/IJCNN.1992.287184
Filename :
287184
Link To Document :
بازگشت