DocumentCode :
288651
Title :
On the synthesis and complexity of feedforward networks
Author :
Sacha, Jaroslaw P. ; Cios, Krzysztof J.
Author_Institution :
Toledo Univ., OH, USA
Volume :
4
fYear :
1994
fDate :
27 Jun-2 Jul 1994
Firstpage :
2185
Abstract :
A method for synthesis of one hidden layer feedforward neural networks approximating functions of several variables to the desired degree of accuracy is presented. The method fully determines the network architecture including the values of all weights. In the first step, the inverse Radon transform is used to decompose a problem of of approximating a function of several variables into several problems of approximating a function of one variable. In the second step, each of the obtained one-dimensional functions is approximated by a sub-network. Then sub-networks are combined to construct the network approximating the original function. The upper bound of the final approximation error ε and errors at each step are estimated. The complexity of the network, or the number of neurons in the hidden layer, is Oε (1/εn), where n is the dimension of the space
Keywords :
Radon transforms; approximation theory; computational complexity; feedforward neural nets; function approximation; functional analysis; approximation error; complexity; decomposition; function approximation; hidden layer feedforward neural networks; inverse Radon transform; network architecture; upper bound; weights; Approximation error; Feedforward neural networks; Indium tin oxide; Integral equations; Multidimensional systems; Network synthesis; Neural networks; Neurons; Transfer functions; Upper bound;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Neural Networks, 1994. IEEE World Congress on Computational Intelligence., 1994 IEEE International Conference on
Conference_Location :
Orlando, FL
Print_ISBN :
0-7803-1901-X
Type :
conf
DOI :
10.1109/ICNN.1994.374555
Filename :
374555
Link To Document :
بازگشت