DocumentCode :
288372
Title :
An update function that speeds up backpropagation learning
Author :
Deredy, W. El ; Branston, N.M.
Author_Institution :
Dept. of Neurological Surg., Inst. of Neurology, London, UK
Volume :
1
fYear :
1994
fDate :
27 Jun-2 Jul 1994
Firstpage :
477
Abstract :
We consider a modification of the backpropagation (BP) learning algorithm in which a linear function, directly proportional to the deviation between target values and actual values at the output, is propagated backwards instead of the original nonlinear function. The new algorithm is tested on the odd/even parity function for orders between 4 and 7 and on high- (180) dimensional data derived from NMR spectroscopy of animal tumours. Results suggest that using the linear function, the network converges faster and is more likely to escape from local minima than when the original BP is used
Keywords :
backpropagation; neural nets; NMR spectroscopy; animal tumours; backpropagation learning; linear function; local minima; neural net; odd/even parity function; update function; Animals; Artificial neural networks; Backpropagation algorithms; Nervous system; Nuclear magnetic resonance; Pattern recognition; Spectroscopy; Supervised learning; Testing; Tumors;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Neural Networks, 1994. IEEE World Congress on Computational Intelligence., 1994 IEEE International Conference on
Conference_Location :
Orlando, FL
Print_ISBN :
0-7803-1901-X
Type :
conf
DOI :
10.1109/ICNN.1994.374209
Filename :
374209
Link To Document :
بازگشت