DocumentCode :
1748797
Title :
Neural network training for varying output node dimension
Author :
Jung, Jae-Byung ; El-Sharkawi, M.A. ; Marks, R.J., II ; Miyamoto, Robert ; Fox, Warren L J ; Anderson, G.M. ; Eggen, C.J.
Author_Institution :
Dept. of Electr. Eng., Washington Univ., Seattle, WA, USA
Volume :
3
fYear :
2001
fDate :
2001
Firstpage :
1733
Abstract :
Considers the problem of neural network supervised learning when the number of output nodes can vary for differing training data. The paper proposes irregular weight updates and learning rate adjustment to compensate for this variation. In order to compensate for possible over training, an a posteriori probability that shows how often the weights associated with each output neuron are updated is obtained from the training data set and is used to evenly distribute the opportunity for weight update to each output neuron. The weight space becomes smoother and the generalization performance is significantly improved
Keywords :
generalisation (artificial intelligence); learning (artificial intelligence); multilayer perceptrons; probability; a posteriori probability; generalization performance; irregular weight updates; learning rate adjustment; neural network supervised learning; neural network training; varying output node dimension; weight space; Computational intelligence; Filling; Laboratories; Multilayer perceptrons; Network topology; Neural networks; Neurons; Physics; Supervised learning; Training data;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Neural Networks, 2001. Proceedings. IJCNN '01. International Joint Conference on
Conference_Location :
Washington, DC
ISSN :
1098-7576
Print_ISBN :
0-7803-7044-9
Type :
conf
DOI :
10.1109/IJCNN.2001.938423
Filename :
938423
Link To Document :
بازگشت