Title :
Implementation of neural networks on massive memory organizations
Author :
Misra, Manavendra ; Prasanna, Viktor K.
Author_Institution :
Dept. of Electr. Eng.-Syst., Univ. of Southern California, Los Angeles, CA, USA
fDate :
7/1/1992 12:00:00 AM
Abstract :
Simulations of artificial neural networks (ANNs) on serial machines have proved to be too slow to be of practical significance. It was realized that parallel machines would have to be used to exploit the inherent parallelism in these models. The SIMD architecture presented has n PEs and n2 memory modules arranged in an n×n array. This massive memory is used to store the weights of the neural network being simulated. It is shown how networks with sparse connectivity among neurons can be simulated in O((n+e)1/2) time, where n is the number of neurons and e the number of interconnections in the network. Preprocessing is carried out on the connection matrix of the sparse network, resulting in data movement that has an optimal asymptotic time complexity and a small constant factor
Keywords :
memory architecture; neural nets; parallel architectures; virtual machines; SIMD architecture; asymptotic time complexity; connection matrix; constant factor; data movement; massive memory organizations; memory modules; neural networks; parallel machines; preprocessing; sparse connectivity; sparse network; Artificial neural networks; Biological neural networks; Biology computing; Computational modeling; Computer networks; Neural networks; Neurons; Parallel machines; Parallel processing; Sparse matrices;
Journal_Title :
Circuits and Systems II: Analog and Digital Signal Processing, IEEE Transactions on