DocumentCode :
3252816
Title :
Training recurrent networks using the extended Kalman filter
Author :
Williams, Ronald J.
Author_Institution :
Coll. of Comput. Sci., Northeastern Univ., Boston, MA, USA
Volume :
4
fYear :
1992
fDate :
7-11 Jun 1992
Firstpage :
241
Abstract :
The author describes some relationships between the extended Kalman filter (EKF) as applied to recurrent net learning and some simpler techniques that are more widely used. In particular, making certain simplifications to the EKF gives rise to an algorithm essentially identical to the real-time recurrent learning (RTRL) algorithm. Since the EKF involves adjusting unit activity in the network, it also provides a principled generalization of the teacher forcing technique. Preliminary simulation experiments on simple finite-state Boolean tasks indicated that the EKF can provide substantial speed-up in number of time steps required for training on such problems when compared with simpler online gradient algorithms. The computational requirements of the EKF are steep, but scale with network size at the same rate as RTRL
Keywords :
Kalman filters; filtering and prediction theory; recurrent neural nets; extended Kalman filter; finite-state Boolean tasks; generalization; real-time recurrent learning; recurrent networks; teacher forcing; Computational efficiency; Computational modeling; Computer networks; Computer science; Current measurement; Educational institutions; Integrated circuit noise; Noise measurement; State estimation; Time measurement;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Neural Networks, 1992. IJCNN., International Joint Conference on
Conference_Location :
Baltimore, MD
Print_ISBN :
0-7803-0559-0
Type :
conf
DOI :
10.1109/IJCNN.1992.227335
Filename :
227335
Link To Document :
بازگشت