DocumentCode
314391
Title
Extensions and enhancements of decoupled extended Kalman filter training
Author
Puskorius, G.V. ; Feldkamp, L.A.
Author_Institution
Sci. Res. Lab., Ford Motor Co., Dearborn, MI, USA
Volume
3
fYear
1997
fDate
9-12 Jun 1997
Firstpage
1879
Abstract
We describe here three useful and practical extensions and enhancements of the decoupled extended Kalman filter (DEKF) neural network weight update procedure, which has served as the backbone for much of our applications-oriented research for the last six years. First, we provide a mechanism that constrains weight values to a pre-specified range during training to allow for fixed-point deployment of trained networks. Second, we examine modifications of DEKF training for alternative cost functions; as an example, we show how to use DEKF training to minimize a measure of relative entropy, rather than mean squared error, for pattern classification problems. Third, we describe an approximation of DEKF training that allows a multiple-output training problem to be treated with single-output training complexity
Keywords
Kalman filters; computational complexity; learning (artificial intelligence); neural nets; DEKF neural network weight update procedure; cost functions; decoupled extended Kalman filter training; fixed-point deployment; multiple-output training problem; pattern classification; relative entropy measure minimization; single-output training complexity; Backpropagation; Cost function; Covariance matrix; Entropy; Equations; Laboratories; Neural networks; Pattern classification; Recurrent neural networks; Spine;
fLanguage
English
Publisher
ieee
Conference_Titel
Neural Networks,1997., International Conference on
Conference_Location
Houston, TX
Print_ISBN
0-7803-4122-8
Type
conf
DOI
10.1109/ICNN.1997.614185
Filename
614185
Link To Document