Title :
Regularization parameter estimation for feedforward neural networks
Author :
Guo, Ping ; Lyu, Michael R. ; Chen, C. L Philip
Author_Institution :
Dept. of Comput. Sci., Beijing Normal Univ., China
fDate :
2/1/2003 12:00:00 AM
Abstract :
Under the framework of the Kullback-Leibler (KL) distance, we show that a particular case of Gaussian probability function for feedforward neural networks (NNs) reduces into the first-order Tikhonov regularizer. The smooth parameter in kernel density estimation plays the role of regularization parameter. Under some approximations, an estimation formula is derived for estimating regularization parameters based on training data sets. The similarity and difference of the obtained results are compared with other work. Experimental results show that the estimation formula works well in sparse and small training sample cases.
Keywords :
feedforward neural nets; learning (artificial intelligence); parameter estimation; probability; Gaussian probability function; Kullback-Leibler distance; feedforward neural networks; first-order Tikhonov regularizer; kernel density estimation; regularization parameter estimation; smooth parameter; training data sets; Computer science; Councils; Density functional theory; Feedforward neural networks; Kernel; Neural networks; Neurons; Parameter estimation; Smoothing methods; Training data;
Journal_Title :
Systems, Man, and Cybernetics, Part B: Cybernetics, IEEE Transactions on
DOI :
10.1109/TSMCB.2003.808176