Title : 
TRTRL: A Localized Resource-Efficient Learning Algorithm for Recurrent Neural Netowrks
         
        
            Author : 
Budik, Danny ; Elhanany, Itamar
         
        
            Author_Institution : 
Univ. of Tennessee, Knoxville
         
        
        
        
        
        
        
            Abstract : 
This paper introduces an efficient, low-complexity online learning algorithm for recurrent neural networks. The approach is based on the real-time recurrent learning (RTRL) algorithm, whereby the sensitivity set of each neuron is reduced to weights associated either with its input or ouput links. As a consequence, storage requirements are reduced from O(N3) to O(N2) and the computational complexity is reduced to O(N2). Despite the radical reduction in resource requirements, it is shown through simulation results that the overall performance degradation is rather minor. Moreover, the scheme lends itself to parallel hardware realization by virtue of the localized property that is inherent to the approach.
         
        
            Keywords : 
computational complexity; learning (artificial intelligence); recurrent neural nets; computational complexity; localized resource-efficient learning algorithm; real-time recurrent learning algorithm; recurrent neural networks; Backpropagation; Computational complexity; Computational modeling; Constraint optimization; Error correction; Neurons; Nonlinear dynamical systems; Nonlinear systems; Recurrent neural networks; System identification; Recurrent neural networks; constraint optimization; real-time recurrent learning (RTRL);
         
        
        
        
            Conference_Titel : 
Circuits and Systems, 2006. MWSCAS '06. 49th IEEE International Midwest Symposium on
         
        
            Conference_Location : 
San Juan
         
        
        
            Print_ISBN : 
1-4244-0172-0
         
        
            Electronic_ISBN : 
1548-3746
         
        
        
            DOI : 
10.1109/MWSCAS.2006.382075