• DocumentCode
    3499322
  • Title

    Optimal output gain algorithm for feed-forward network training

  • Author

    Aswathappa, Babu Hemanth Kumar ; Manry, M.T. ; Rawat, Rohit

  • Author_Institution
    Intel Corp., Chandler, AZ, USA
  • fYear
    2011
  • fDate
    July 31 2011-Aug. 5 2011
  • Firstpage
    2609
  • Lastpage
    2616
  • Abstract
    A batch training algorithm for feed-forward networks is proposed which uses Newton´s method to estimate a vector of optimal scaling factors for output errors in the network. Using this vector, backpropagation is used to modify weights feeding into the hidden units. Linear equations are then solved for the network´s output weights. Elements of the new method´s Gauss-Newton Hessian matrix are shown to be weighted sums of elements from the total network´s Hessian. The effect of output transformation on training a feed-forward network is reviewed and explained, using the concept of equivalent networks. In several examples, the new method performs better than backpropagation and conjugate gradient, with similar numbers of required multiplies. The method performs almost as well as Levenberg-Marquardt, with several orders of magnitude fewer multiplies due to the small size of the new method´s Hessian.
  • Keywords
    Hessian matrices; Newton method; learning (artificial intelligence); multilayer perceptrons; Gauss-Newton Hessian matrix; Levenberg-Marquardt algorithm; Newton method; batch training algorithm; equivalent networks concept; feedforward network training; linear equation; optimal output gain algorithm; optimal scaling factor; Backpropagation; Equations; Mathematical model; Matrix decomposition; Newton method; Training; Vectors;
  • fLanguage
    English
  • Publisher
    ieee
  • Conference_Titel
    Neural Networks (IJCNN), The 2011 International Joint Conference on
  • Conference_Location
    San Jose, CA
  • ISSN
    2161-4393
  • Print_ISBN
    978-1-4244-9635-8
  • Type

    conf

  • DOI
    10.1109/IJCNN.2011.6033559
  • Filename
    6033559