DocumentCode :
1346726
Title :
New results on recurrent network training: unifying the algorithms and accelerating convergence
Author :
Atiya, Amir F. ; Parlos, Alexander G.
Author_Institution :
Dept. of Electr. Eng., California Inst. of Technol., Pasadena, CA, USA
Volume :
11
Issue :
3
fYear :
2000
fDate :
5/1/2000 12:00:00 AM
Firstpage :
697
Lastpage :
709
Abstract :
How to efficiently train recurrent networks remains a challenging and active research topic. Most of the proposed training approaches are based on computational ways to efficiently obtain the gradient of the error function, and can be generally grouped into five major groups. In this study we present a derivation that unifies these approaches. We demonstrate that the approaches are only five different ways of solving a particular matrix equation. The second goal of this paper is develop a new algorithm based on the insights gained from the novel formulation. The new algorithm, which is based on approximating the error gradient, has lower computational complexity in computing the weight update than the competing techniques for most typical problems. In addition, it reaches the error minimum in a much smaller number of iterations. A desirable characteristic of recurrent network training algorithms is to be able to update the weights in an online fashion. We have also developed an online version of the proposed algorithm, that is based on updating the error gradient approximation in a recursive manner
Keywords :
computational complexity; convergence; learning (artificial intelligence); recurrent neural nets; accelerated convergence; algorithm unification; computational complexity; error function gradient; error gradient approximation; error minimum; matrix equation; online weight updating; recurrent network training algorithms; recurrent neural network training; Acceleration; Approximation algorithms; Backpropagation algorithms; Computational complexity; Convergence; Differential equations; Error correction; Nonlinear dynamical systems; Optimal control; Time factors;
fLanguage :
English
Journal_Title :
Neural Networks, IEEE Transactions on
Publisher :
ieee
ISSN :
1045-9227
Type :
jour
DOI :
10.1109/72.846741
Filename :
846741
Link To Document :
بازگشت