Title :
A Fast and Scalable Recurrent Neural Network Based on Stochastic Meta Descent
Author :
Liu, Zhenzhen ; Elhanany, Itamar
Author_Institution :
Electr. Eng. & Comput. Sci., Tennessee Univ., Knoxville, TN
Abstract :
This brief presents an efficient and scalable online learning algorithm for recurrent neural networks (RNNs). The approach is based on the real-time recurrent learning (RTRL) algorithm, whereby the sensitivity set of each neuron is reduced to weights associated with either its input or output links. This yields a reduced storage and computational complexity of O(N2). Stochastic meta descent (SMD), an adaptive step size scheme for stochastic gradient-descent problems, is employed as means of incorporating curvature information in order to substantially accelerate the learning process. We also introduce a clustered version of our algorithm to further improve its scalability attributes. Despite the dramatic reduction in resource requirements, it is shown through simulation results that the approach outperforms regular RTRL by almost an order of magnitude. Moreover, the scheme lends itself to parallel hardware realization by virtue of the localized property that is inherent to the learning framework.
Keywords :
computational complexity; gradient methods; learning (artificial intelligence); recurrent neural nets; stochastic processes; computational complexity; online learning algorithm; real-time recurrent learning algorithm; resource requirements; scalable recurrent neural network; stochastic gradient-descent problems; stochastic meta descent; Constrained optimization; real-time recurrent learning (RTRL); recurrent neural networks (RNNs); Algorithms; Artificial Intelligence; Computer Simulation; Models, Statistical; Neural Networks (Computer); Pattern Recognition, Automated; Stochastic Processes;
Journal_Title :
Neural Networks, IEEE Transactions on
DOI :
10.1109/TNN.2008.2000838