DocumentCode
3499322
Title
Optimal output gain algorithm for feed-forward network training
Author
Aswathappa, Babu Hemanth Kumar ; Manry, M.T. ; Rawat, Rohit
Author_Institution
Intel Corp., Chandler, AZ, USA
fYear
2011
fDate
July 31 2011-Aug. 5 2011
Firstpage
2609
Lastpage
2616
Abstract
A batch training algorithm for feed-forward networks is proposed which uses Newton´s method to estimate a vector of optimal scaling factors for output errors in the network. Using this vector, backpropagation is used to modify weights feeding into the hidden units. Linear equations are then solved for the network´s output weights. Elements of the new method´s Gauss-Newton Hessian matrix are shown to be weighted sums of elements from the total network´s Hessian. The effect of output transformation on training a feed-forward network is reviewed and explained, using the concept of equivalent networks. In several examples, the new method performs better than backpropagation and conjugate gradient, with similar numbers of required multiplies. The method performs almost as well as Levenberg-Marquardt, with several orders of magnitude fewer multiplies due to the small size of the new method´s Hessian.
Keywords
Hessian matrices; Newton method; learning (artificial intelligence); multilayer perceptrons; Gauss-Newton Hessian matrix; Levenberg-Marquardt algorithm; Newton method; batch training algorithm; equivalent networks concept; feedforward network training; linear equation; optimal output gain algorithm; optimal scaling factor; Backpropagation; Equations; Mathematical model; Matrix decomposition; Newton method; Training; Vectors;
fLanguage
English
Publisher
ieee
Conference_Titel
Neural Networks (IJCNN), The 2011 International Joint Conference on
Conference_Location
San Jose, CA
ISSN
2161-4393
Print_ISBN
978-1-4244-9635-8
Type
conf
DOI
10.1109/IJCNN.2011.6033559
Filename
6033559
Link To Document