Title :
Reducing the number of multiplies in backpropagation
Author :
Boonyanit, Kan ; Peterson, Allen M.
Author_Institution :
LSI Logic Corp., Milpitas, CA, USA
fDate :
27 Jun-2 Jul 1994
Abstract :
There have been many algorithms to speed up the learning time of backpropagation. However, most of them do not take into consideration the amount of hardware required to implement the algorithm. Without suitable hardware implementation, the real promise of neural network applications will be difficult to achieve. Since multiply dominates computation and is expensive in hardware, this paper proposes a method to reduce the number of multiplies in the backward path of backpropagation algorithm by setting some neuron errors to zero. It proves the convergence theorem by the general Robbins-Monro process, a stochastic approximation process
Keywords :
approximation theory; backpropagation; convergence of numerical methods; neural nets; Robbins-Monro process; backpropagation; backward path; convergence theorem; learning time; multiply reduction; neural network; neuron errors; stochastic approximation; Backpropagation algorithms; Convergence; Data flow computing; Large scale integration; Logic; Neural network hardware; Neural networks; Neurons; Noise reduction; Stochastic processes;
Conference_Titel :
Neural Networks, 1994. IEEE World Congress on Computational Intelligence., 1994 IEEE International Conference on
Conference_Location :
Orlando, FL
Print_ISBN :
0-7803-1901-X
DOI :
10.1109/ICNN.1994.374133