DocumentCode
2199047
Title
Parallel and separable recursive Levenberg-Marquardt training algorithm
Author
Asirvadam, V.S. ; McLoone, S.F. ; Irwin, G.W.
Author_Institution
Sch. of Electr. & Electron. Eng., Queen´´s Univ., Belfast, UK
fYear
2002
fDate
2002
Firstpage
129
Lastpage
138
Abstract
A novel decomposed recursive Levenberg Marquardt (RLM) algorithm is derived for the training of feedforward neural networks. By neglecting interneuron weight correlations the recently proposed RLM training algorithm can be decomposed at neuron level enabling weights to be updated in an efficient parallel manner. A separable least squares implementation of decomposed RLM is also introduced. Experiment results for two nonlinear time series problems demonstrate the superiority of the new training algorithms.
Keywords
feedforward neural nets; learning (artificial intelligence); least squares approximations; nonlinear estimation; parallel algorithms; recursive estimation; time series; Levenberg-Marquardt training algorithm; decomposed RLM; feedforward neural networks; neuron level; nonlinear time series problems; parallel algorithm; separable least squares implementation; separable recursive training algorithm; weight updating; Backpropagation algorithms; Convergence; Cost function; Feedforward neural networks; Least squares methods; Neural networks; Neurons; Partitioning algorithms; Resonance light scattering; Training data;
fLanguage
English
Publisher
ieee
Conference_Titel
Neural Networks for Signal Processing, 2002. Proceedings of the 2002 12th IEEE Workshop on
Print_ISBN
0-7803-7616-1
Type
conf
DOI
10.1109/NNSP.2002.1030024
Filename
1030024
Link To Document