DocumentCode
353221
Title
Weight groupings in the training of recurrent networks
Author
Chan, Lai-Wan ; Szeto, Chi-Cheong
Author_Institution
Comput. Sci. & Eng. Dept., Chinese Univ. of Hong Kong, Shatin, Hong Kong
Volume
3
fYear
2000
fDate
2000
Firstpage
21
Abstract
We use the block-diagonal matrix to approximate the Hessian matrix in the Levenberg Marquardt method for the training of recurrent neural networks. Substantial improvement of the training time over the original Levenberg Marquardt method is observed without degrading the generalization ability. Three weight grouping methods, correlation blocks, k-unit blocks and layer blocks were investigated and compared. Their computational complexity, approximation ability, and training time are analyzed
Keywords
Hessian matrices; computational complexity; generalisation (artificial intelligence); learning (artificial intelligence); recurrent neural nets; Hessian matrix; Levenberg Marquardt method; approximation; block-diagonal matrix; computational complexity; correlation blocks; generalization; learning time; recurrent neural networks; weight grouping; Approximation methods; Computer science; Degradation; Equations; Intelligent networks; Jacobian matrices; Learning systems; Matrix decomposition; Neurons; Recurrent neural networks;
fLanguage
English
Publisher
ieee
Conference_Titel
Neural Networks, 2000. IJCNN 2000, Proceedings of the IEEE-INNS-ENNS International Joint Conference on
Conference_Location
Como
ISSN
1098-7576
Print_ISBN
0-7695-0619-4
Type
conf
DOI
10.1109/IJCNN.2000.861275
Filename
861275
Link To Document