DocumentCode :
1126565
Title :
Regularization parameter estimation for feedforward neural networks
Author :
Guo, Ping ; Lyu, Michael R. ; Chen, C. L Philip
Author_Institution :
Dept. of Comput. Sci., Beijing Normal Univ., China
Volume :
33
Issue :
1
fYear :
2003
fDate :
2/1/2003 12:00:00 AM
Firstpage :
35
Lastpage :
44
Abstract :
Under the framework of the Kullback-Leibler (KL) distance, we show that a particular case of Gaussian probability function for feedforward neural networks (NNs) reduces into the first-order Tikhonov regularizer. The smooth parameter in kernel density estimation plays the role of regularization parameter. Under some approximations, an estimation formula is derived for estimating regularization parameters based on training data sets. The similarity and difference of the obtained results are compared with other work. Experimental results show that the estimation formula works well in sparse and small training sample cases.
Keywords :
feedforward neural nets; learning (artificial intelligence); parameter estimation; probability; Gaussian probability function; Kullback-Leibler distance; feedforward neural networks; first-order Tikhonov regularizer; kernel density estimation; regularization parameter estimation; smooth parameter; training data sets; Computer science; Councils; Density functional theory; Feedforward neural networks; Kernel; Neural networks; Neurons; Parameter estimation; Smoothing methods; Training data;
fLanguage :
English
Journal_Title :
Systems, Man, and Cybernetics, Part B: Cybernetics, IEEE Transactions on
Publisher :
ieee
ISSN :
1083-4419
Type :
jour
DOI :
10.1109/TSMCB.2003.808176
Filename :
1167352
Link To Document :
بازگشت