DocumentCode :
303249
Title :
Second differentials in arbitrary feedforward neural networks
Author :
Rossi, Fabrice
Author_Institution :
Thomson-CSF, Bagneux, France
Volume :
1
fYear :
1996
fDate :
3-6 Jun 1996
Firstpage :
418
Abstract :
We extend here a general mathematical model for feedforward neural networks. Such a network is represented as a vectorial function f of two variables, x (the input of the network) and w (the weight vector). We have already shown that the differential of f can be computed with an extended back-propagation algorithm as well as with a direct method. In this paper, we show that the second differentials of f can also be computed with several different algorithms. Evaluating the theoretical complexities of these methods allow one to choose the fastest algorithm for a particular architecture. This will allow us to handle arbitrary feedforward neural network learning with the help of recent training and analysis techniques based on the Hessian matrix of the error
Keywords :
Hessian matrices; feedforward neural nets; learning (artificial intelligence); Hessian matrix; extended backpropagation algorithm; fastest algorithm; feedforward neural networks; second differentials; vectorial function; Communication system control; Computer architecture; Computer networks; Electronic mail; Feedforward neural networks; Feedforward systems; Intelligent networks; Mathematical model; Neural networks; Neurons;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Neural Networks, 1996., IEEE International Conference on
Conference_Location :
Washington, DC
Print_ISBN :
0-7803-3210-5
Type :
conf
DOI :
10.1109/ICNN.1996.548929
Filename :
548929
Link To Document :
بازگشت