Title :
Fine-Grain Parallelization of Recurrent Neural Networks Training
Author :
Turchenko, Volodymyr
Author_Institution :
Dept. of Inf. Comput. Syst. & Control, Ternopil State Econ. Univ., Ternopil
fDate :
Feb. 28 2006-March 4 2006
Abstract :
An approach to development of fine-grain parallel algorithm of artificial neural network training using parallelization of computational operations of each elementary neuron is presented in this paper. A training algorithm of back error propagation is described and parallel section of the algorithm is developed. The results of experimental research of the parallel algorithm are given using analysis of parallelization speedup and efficiency on parallel computer Origin 300.
Keywords :
backpropagation; parallel algorithms; parallel machines; parallel processing; recurrent neural nets; Origin 300 parallel computer; artificial neural network training; back error propagation; computational parallelization; fine-grain parallel algorithm; fine-grain parallelization; parallelization efficiency; parallelization speedup; recurrent neural networks training; training algorithm; Artificial neural networks; Computer errors; Computer networks; Concurrent computing; Hardware; Neural networks; Neurons; Parallel algorithms; Parallel processing; Recurrent neural networks; fine-grain parallelization; parallel computer; recurrent neural network;
Conference_Titel :
Modern Problems of Radio Engineering, Telecommunications, and Computer Science, 2006. TCSET 2006. International Conference
Conference_Location :
Lviv-Slavsko
Print_ISBN :
966-553-507-2
DOI :
10.1109/TCSET.2006.4404497