Title :
On derivation of stagewise second-order backpropagation by invariant imbedding for multi-stage neural-network learning
Author :
Mizutani, Eiji ; Dreyfus, Stuart
Author_Institution :
Tsing Hua Univ., Hsinchu
Abstract :
We present a simple, intuitive argument based on "invariant imbedding" in the spirit of dynamic programming to derive a stagewise second-order backpropagation (BP) algorithm. The method evaluates the Hessian matrix of a general objective function efficiently by exploiting the multistage structure embedded in a given neural-network model such as a multilayer perceptron (MLP). In consequence, for instance, our stagewise BP can compute the full Hessian matrix "faster" than the standard method that evaluates the Gauss-Newton Hessian matrix alone by rank updates in nonlinear least squares learning. Through our derivation, we also show how the procedure serves to develop advanced learning algorithms; in particular, we explain how the introduction of "stage costs" leads to alternative systematic implementations of multi-task learning and weight decay.
Keywords :
Hessian matrices; backpropagation; dynamic programming; multilayer perceptrons; Gauss-Newton Hessian matrix; backpropagation algorithm; dynamic programming; invariant imbedding; learning algorithms; multilayer perceptron; multistage neural network learning; multitask learning; neural network model; nonlinear least squares learning; stagewise second-order backpropagation; Backpropagation algorithms; Costs; Dynamic programming; Equations; Gaussian processes; Least squares methods; Multi-layer neural network; Multilayer perceptrons; Neural networks; Optimal control;
Conference_Titel :
Neural Networks, 2006. IJCNN '06. International Joint Conference on
Conference_Location :
Vancouver, BC
Print_ISBN :
0-7803-9490-9
DOI :
10.1109/IJCNN.2006.247151