Title :
Universal approximation using incremental constructive feedforward networks with random hidden nodes
Author :
Guang-Bin Huang ; Lei Chen ; Chee-Kheong Siew
Author_Institution :
Sch. of Electr. & Electron. Eng., Nanyang Technol. Univ., Singapore, Singapore
fDate :
7/1/2006 12:00:00 AM
Abstract :
According to conventional neural network theories, single-hidden-layer feedforward networks (SLFNs) with additive or radial basis function (RBF) hidden nodes are universal approximators when all the parameters of the networks are allowed adjustable. However, as observed in most neural network implementations, tuning all the parameters of the networks may cause learning complicated and inefficient, and it may be difficult to train networks with nondifferential activation functions such as threshold networks. Unlike conventional neural network theories, this paper proves in an incremental constructive method that in order to let SLFNs work as universal approximators, one may simply randomly choose hidden nodes and then only need to adjust the output weights linking the hidden layer and the output layer. In such SLFNs implementations, the activation functions for additive nodes can be any bounded nonconstant piecewise continuous functions g:R→R and the activation functions for RBF nodes can be any integrable piecewise continuous functions g:R→R and ∫Rg(x)dx≠0. The proposed incremental method is efficient not only for SFLNs with continuous (including nondifferentiable) activation functions but also for SLFNs with piecewise continuous (such as threshold) activation functions. Compared to other popular methods such a new network is fully automatic and users need not intervene the learning process by manually tuning control parameters.
Keywords :
learning (artificial intelligence); radial basis function networks; transfer functions; additive nodes; bounded nonconstant piecewise continuous functions; incremental constructive feedforward networks; integrable piecewise continuous functions; neural network theories; nondifferential activation functions; radial basis function hidden nodes; random hidden nodes; single-hidden-layer feedforward networks; threshold networks; universal approximation; Artificial neural networks; Automatic control; Feedforward neural networks; Function approximation; Helium; Joining processes; Machine learning; Neural networks; Radial basis function networks; Support vector machines; Ensemble; feedforward network; incremental extreme learning machine; radial basis function; random hidden nodes; support vector machine; threshold network; universal approximation;
Journal_Title :
Neural Networks, IEEE Transactions on
DOI :
10.1109/TNN.2006.875977