Title :
Neural networks as function approximators: teaching a neural network to multiply
Author :
Vaccari, David A. ; Wojciechowski, Edward
Author_Institution :
Stevens Inst. of Technol., Hoboken, NJ, USA
fDate :
27 Jun-2 Jul 1994
Abstract :
Artificial neural networks (ANNs) were first proposed, by Hecht-Nieisen (1987), as multivariate function approximators based on Kolmogorov´s theorem. Since then, several researchers have proven that multilayer ANNs, with an arbitrary squashing function in the hidden layer, can approximate any multivariate function to any degree of accuracy. Based on these results, researchers have attempted to train backpropagation networks to realize arbitrary functions. Although their results are encouraging, this technique has many shortcomings and may lead to an inappropriate response by the network. In this paper, the authors present an alternative neural network architecture, based on cascaded univariate function approximators, which can be trained to multiply two real numbers and may be used to realize arbitrary multivariate function mappings
Keywords :
backpropagation; function approximation; neural net architecture; neural nets; Kolmogorov´s theorem; backpropagation networks; cascaded univariate function approximators; hidden layer; multivariate function approximators; multivariate function mappings; neural networks; squashing function; Aerospace electronics; Artificial neural networks; Backpropagation; Education; Error correction; Function approximation; Multi-layer neural network; Network topology; Neural networks; Shape;
Conference_Titel :
Neural Networks, 1994. IEEE World Congress on Computational Intelligence., 1994 IEEE International Conference on
Conference_Location :
Orlando, FL
Print_ISBN :
0-7803-1901-X
DOI :
10.1109/ICNN.1994.374561