Title :
Parallel implementation of a recursive least squares neural network training method on the Intel iPSC/2
Author :
Steck, James Edward ; McMillin, Bruce M. ; Krishnamurthy, K. ; Ashouri, M. Reza ; Leininger, Gary G.
Abstract :
An algorithm based on the Marquardt-Levenberg least-square optimization method has been shown by S. Kollias and D. Anastassiou (IEEE Trans. on Circuits Syst. vol.36, no.8, p.1092-101, Aug. 1989) to be a much more efficient training method than gradient descent, when applied to some small feedforward neural networks. Yet, for many applications, the increase in computational complexity of the method outweighs any gain in learning rate obtained over current training methods. However, the least-squares method can be more efficiently implemented on parallel architectures than standard methods. This is demonstrated by comparing computation times and learning rates for the least-squares method implemented on 1, 2, 4, 8, and 16 processors on an Intel iPSC/2 multicomputer. Two applications which demonstrate the faster real-time learning rate of the last-squares method over than of gradient descent are given
Keywords :
convergence; learning systems; least squares approximations; neural nets; optimisation; parallel processing; Intel iPSC/2; Marquardt-Levenberg least-square optimization method; computation times; learning rates; parallel architectures; recursive least squares neural network training; supervised learning;
Conference_Titel :
Neural Networks, 1990., 1990 IJCNN International Joint Conference on
Conference_Location :
San Diego, CA, USA
DOI :
10.1109/IJCNN.1990.137641