Title :
Effects of moving the center´s in an RBF network
Author :
Panchapakesan, Chitra ; Palaniswami, Marimuthu ; Ralph, Daniel ; Manzie, Chris
Author_Institution :
Dept. of Electr. & Electron. Eng., Univ. of Melbourne, Vic., Australia
fDate :
11/1/2002 12:00:00 AM
Abstract :
In radial basis function (RBF) networks, placement of centers is said to have a significant effect on the performance of the network. Supervised learning of center locations in some applications show that they are superior to the networks whose centers are located using unsupervised methods. But such networks can take the same training time as that of sigmoid networks. The increased time needed for supervised learning offsets the training time of regular RBF networks. One way to overcome this may be to train the network with a set of centers selected by unsupervised methods and then to fine tune the locations of centers. This can be done by first evaluating whether moving the centers would decrease the error and then, depending on the required level of accuracy, changing the center locations. This paper provides new results on bounds for the gradient and Hessian of the error considered first as a function of the independent set of parameters, namely the centers, widths, and weights; and then as a function of centers and widths where the linear weights are now functions of the basis function parameters for networks of fixed size. Moreover, bounds for the Hessian are also provided along a line beginning at the initial set of parameters. Using these bounds, it is possible to estimate how much one can reduce the error by changing the centers. Further to that, a step size can be specified to achieve a guaranteed, amount of reduction in error.
Keywords :
Hessian matrices; learning (artificial intelligence); neural net architecture; radial basis function networks; Hessian matrices; RBF network; center locations; error; gradient methods; intelligent networks; network performance; neural network architecture; nonlinear estimation; radial basis function networks; sigmoid networks; supervised learning; training time; unsupervised learning; Engineering management; Function approximation; Gradient methods; Intelligent networks; Interpolation; Learning systems; Pattern recognition; Radial basis function networks; Shape; Supervised learning;
Journal_Title :
Neural Networks, IEEE Transactions on
DOI :
10.1109/TNN.2002.804286