DocumentCode
801140
Title
Sensitivity analysis of single hidden-layer neural networks with threshold functions
Author
Oh, Sang-Hoon ; Lee, Youngjik
Author_Institution
Res. Dept., Electron. & Telecommun. Res. Inst., Daejeon, South Korea
Volume
6
Issue
4
fYear
1995
fDate
7/1/1995 12:00:00 AM
Firstpage
1005
Lastpage
1007
Abstract
An important consideration when applying neural networks to pattern recognition is the sensitivity to weight perturbation or to input errors. In this paper, we analyze the sensitivity of single hidden-layer networks with threshold functions. In a case of weight perturbation or input errors, the probability of inversion error for an output neuron is derived as a function of the trained weights, the input pattern, and the variance of weight perturbation or the bit error probability of the input pattern. The derived results are verified with a simulation of the Madaline recognizing handwritten digits. The result shows that the sensitivity of trained networks is far different from that of networks with random weights
Keywords
character recognition; error statistics; neural nets; sensitivity analysis; Madaline; bit error probability; handwritten digit recognition; input errors; inversion error probability; sensitivity analysis; single hidden-layer neural networks; threshold functions; weight perturbation; Approximation methods; Degradation; Error probability; Handwriting recognition; Joining processes; Neural network hardware; Neural networks; Neurons; Pattern recognition; Sensitivity analysis;
fLanguage
English
Journal_Title
Neural Networks, IEEE Transactions on
Publisher
ieee
ISSN
1045-9227
Type
jour
DOI
10.1109/72.392264
Filename
392264
Link To Document