Title :
Modified Kalman filter based method for training state-recurrent multilayer perceptrons
Author :
Erdogmus, Deniz ; Sanchez, Justin C. ; Principe, Jose C.
Author_Institution :
Dept. of Electr. & Comput. Eng., Florida Univ., Gainesville, FL, USA
Abstract :
Kalman filter based training algorithms for recurrent neural networks provide a clever alternative to the standard backpropagation in time. However, these algorithms do not take into account the optimization of the hidden state variables of the recurrent network. In addition, their formulation requires Jacobian evaluations over the entire network, adding to their computational complexity. We propose a spatial-temporal extended Kalman filter algorithm for training recurrent neural network weights and internal states. This new formulation also reduces the computational complexity of Jacobian evaluations drastically by decoupling the gradients of each layer. Monte Carlo comparisons with backpropagation through time point out the robust and fast convergence of the algorithm.
Keywords :
Kalman filters; Monte Carlo methods; backpropagation; computational complexity; filtering theory; multilayer perceptrons; nonlinear filters; recurrent neural nets; Jacobian evaluations; Kalman filter based training algorithms; Monte Carlo method; algorithm convergence; backpropagation in time; computational complexity reduction; hidden state variables optimization; modified Kalman filter; recurrent neural network weights; recurrent neural networks; spatial-temporal extended Kalman filter algorithm; state-recurrent multilayer perceptrons; Backpropagation algorithms; Computational complexity; Convergence; Jacobian matrices; Kalman filters; Multilayer perceptrons; Neural engineering; Neural networks; Recurrent neural networks; Signal processing algorithms;
Conference_Titel :
Neural Networks for Signal Processing, 2002. Proceedings of the 2002 12th IEEE Workshop on
Print_ISBN :
0-7803-7616-1
DOI :
10.1109/NNSP.2002.1030033