DocumentCode :
2260476
Title :
Training feedforward neural networks using orthogonal iteration of the Hessian eigenvectors
Author :
Hunter, Andrew
Author_Institution :
Dept. of Comput. & Eng. Technol., Sunderland Univ., UK
Volume :
2
fYear :
2000
fDate :
2000
Firstpage :
173
Abstract :
The paper describes a training algorithm for multilayer perceptrons. It has scalable memory requirements, which may range from O(W) to O(W2), although in practice the useful range is limited to lower complexity levels. The algorithm is based upon a novel iterative estimation of the principal eigensubspace of the Hessian, together with a quadratic step estimation procedure. It is shown that the new algorithm has convergence time comparable to conjugate gradient descent, and may be preferable if early stopping is used as it converges more quickly during the initial phases. Results of experiments to confirm the algorithm´s performance are presented
Keywords :
Hessian matrices; computational complexity; convergence; eigenvalues and eigenfunctions; feedforward neural nets; iterative methods; learning (artificial intelligence); multilayer perceptrons; Hessian eigenvectors; convergence time; feedforward neural network training; iterative estimation; multilayer perceptrons; orthogonal iteration; principal eigensubspace; quadratic step estimation procedure; scalable memory requirements; Approximation algorithms; Computer networks; Convergence; Eigenvalues and eigenfunctions; Feedforward neural networks; Gradient methods; Multi-layer neural network; Neural networks; Shape;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Neural Networks, 2000. IJCNN 2000, Proceedings of the IEEE-INNS-ENNS International Joint Conference on
Conference_Location :
Como
ISSN :
1098-7576
Print_ISBN :
0-7695-0619-4
Type :
conf
DOI :
10.1109/IJCNN.2000.857893
Filename :
857893
Link To Document :
بازگشت