Title :
On the required size of multilayer networks for implementing real-valued functions
Author_Institution :
Dept. of Comput. Eng., Cairo Univ., Egypt
fDate :
27 Jun-2 Jul 1994
Abstract :
One of the important theoretical issues studied by neural network researchers is how large should the network be to realize an arbitrary set of training patterns. Baum (1988) considered the case of two-class classification problems, where the input vectors are in general position. By general position the author means that no D+1 vectors lie on a (D-1)dimensional hyperplane. He proved that [M/D] hidden nodes are both necessary and sufficient for implementing any arbitrary dichotomy, where M denotes the number of examples, D denotes the dimension of the pattern vectors, and [x] means the smallest integer ⩾x. Buang and Huang (1991) and Sartori and Antsaklis (1991) proved that for the case that the general position condition does not hold, M-1 hidden nodes are sufficient for implementing analog mappings. In this paper the author considers analog mappings (real-valued input vectors and real-valued scalar outputs), and assumes the general position condition. It is proved that 2[M/D] hidden nodes are sufficient for implementing arbitrary mappings
Keywords :
learning (artificial intelligence); multilayer perceptrons; analog mappings; arbitrary mappings; general position condition; hidden nodes; multilayer networks; real-valued functions; real-valued input vectors; real-valued scalar outputs; two-class classification problems; Computer networks; Multi-layer neural network; Neural networks; Nonhomogeneous media; Transfer functions;
Conference_Titel :
Neural Networks, 1994. IEEE World Congress on Computational Intelligence., 1994 IEEE International Conference on
Conference_Location :
Orlando, FL
Print_ISBN :
0-7803-1901-X
DOI :
10.1109/ICNN.1994.374500