Title :
Reduced HyperBF Networks: Regularization by Explicit Complexity Reduction and Scaled Rprop-Based Training
Author :
Mahdi, Rami N. ; Rouchka, Eric Christian
Author_Institution :
Dept. of Genetic Med., Weill Cornell Med. Coll., New York, NY, USA
fDate :
5/1/2011 12:00:00 AM
Abstract :
Hyper basis function (HyperBF) networks are generalized radial basis function neural networks (where the activation function is a radial function of a weighted distance. Such generalization provides HyperBF networks with high capacity to learn complex functions, which in turn make them susceptible to overfitting and poor generalization. Moreover, training a HyperBF network demands the weights, centers, and local scaling factors to be optimized simultaneously. In the case of a relatively large dataset with a large network structure, such optimization becomes computationally challenging. In this paper, a new regularization method that performs soft local dimension reduction in addition to weight decay is proposed. The regularized HyperBF network is shown to provide classification accuracy competitive to a support vector machine while requiring a significantly smaller network structure. Furthermore, a practical training to construct HyperBF networks is presented. Hierarchal clustering is used to initialize neurons followed by a gradient optimization using a scaled version of the Rprop algorithm with a localized partial backtracking step. Experimental results on seven datasets show that the proposed training provides faster and smoother convergence than the regular Rprop algorithm.
Keywords :
backpropagation; backtracking; computational complexity; gradient methods; optimisation; pattern classification; pattern clustering; radial basis function networks; support vector machines; transfer functions; Rprop algorithm; activation function; classification accuracy; complex functions; explicit complexity reduction; generalized radial basis function neural networks; gradient optimization; hierarchal clustering; hyper basis function networks; localized partial backtracking step; network structure; neurons; radial function; reduced HyperBF networks; regularization; regularized HyperBF network; scaled Rprop-based training; scaled version; scaling factors; soft local dimension reduction; support vector machine; weight decay; weighted distance; Complexity theory; Neurons; Optimization; Radial basis function networks; Shape; Training; Training data; Bridge regression; HyperBF; generalized RBF; localized dimension reduction; reduced HyperBF; regularized HyperBF; weight decay; Algorithms; Artificial Intelligence; Automatic Data Processing; Computer Simulation; Humans; Mathematical Concepts; Neural Networks (Computer); Normal Distribution; Pattern Recognition, Automated; Software Design;
Journal_Title :
Neural Networks, IEEE Transactions on
DOI :
10.1109/TNN.2011.2109736