DocumentCode :
3264100
Title :
Self-adaptive learning rates in backpropagation algorithm improve its function approximation performance
Author :
Bhattacharya, U. ; Parui, S.K.
Author_Institution :
Comput. Vision & Pattern Recognition Unit, Indian Stat. Inst., Calcutta, India
Volume :
5
fYear :
1995
fDate :
Nov/Dec 1995
Firstpage :
2784
Abstract :
The backpropagation algorithm helps a multilayer perceptron to learn to map a set of inputs to a set of outputs. But often its function approximation performance is not impressive. In this paper the authors demonstrate that self-adaptation of the learning rate of the backpropagation algorithm helps in improving the approximation of a function. The modified backpropagation algorithm with self-adaptive learning rates is based on a combination of two updating rules-one for updating the connection weights and the other for updating the learning rate. The method for learning rate updating implements the gradient descent principle on the error surface. Simulation results with astrophysical data are presented
Keywords :
backpropagation; function approximation; multilayer perceptrons; self-adjusting systems; backpropagation algorithm; connection weights; error surface; function approximation performance; gradient descent principle; multilayer perceptron; self-adaptive learning rates; Approximation algorithms; Backpropagation algorithms; Computational modeling; Computer errors; Computer vision; Function approximation; Joining processes; Multilayer perceptrons; Neural networks; Pattern recognition;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Neural Networks, 1995. Proceedings., IEEE International Conference on
Conference_Location :
Perth, WA
Print_ISBN :
0-7803-2768-3
Type :
conf
DOI :
10.1109/ICNN.1995.488172
Filename :
488172
Link To Document :
بازگشت