Title :
A local linearized least squares algorithm for training feedforward neural networks
Author :
Stan, Octavian ; Kamen, Edward
Author_Institution :
Sch. of Electr. & Comput. Eng., Georgia Inst. of Technol., Atlanta, GA, USA
fDate :
3/1/2000 12:00:00 AM
Abstract :
In training the weights of a feedforward neural network, it is well known that the global extended Kalman filter (GEKF) algorithm has much better performance than the popular gradient descent with error backpropagation in terms of convergence and quality of solution. However, the GEKF is very computationally intensive, which has led to the development of efficient algorithms such as the multiple extended Kalman algorithm (MEKA) and the decoupled extended Kalman filter algorithm (DEKF), that are based on dimensional reduction and/or partitioning of the global problem. In this paper we present a new training algorithm, called local linearized least squares (LLLS), that is based on viewing the local system identification subproblems at the neuron level as recursive linearized least squares problems. The objective function of the least squares problems for each neuron is the sum of the squares of the linearized backpropagated error signals. The new algorithm is shown to give better convergence results for three benchmark problems in comparison to MEKA, and in comparison to DEKF for highly coupled applications. The performance of the LLLS algorithm approaches that of the GEKF algorithm in the experiments
Keywords :
Kalman filters; backpropagation; computational complexity; convergence; feedforward neural nets; filtering theory; least squares approximations; linearisation techniques; DEKF; GEKF algorithm; LLLS; MEKA; convergence; decoupled extended Kalman filter algorithm; dimensional reduction; efficient algorithms; feedforward neural network training; global extended Kalman filter; linearized backpropagated error signals; local linearized least squares algorithm; local linearized least-squares algorithm; local system identification subproblems; multiple extended Kalman algorithm; partitioning; recursive linearized least-squares problems; solution quality; weight training; Backpropagation algorithms; Differential equations; Feedforward neural networks; Kalman filters; Least squares methods; Multilayer perceptrons; Neural networks; Neurons; Partitioning algorithms; Testing;
Journal_Title :
Neural Networks, IEEE Transactions on