DocumentCode :
2089787
Title :
Error analysis of quantized weights for feedforward neural networks (FNN)
Author :
Wu, Duanpei ; Gowdy, J.N.
Author_Institution :
Dept. of Electr. & Comput. Eng., Clemson Univ., SC, USA
fYear :
1994
fDate :
10-13 Apr 1994
Firstpage :
475
Lastpage :
479
Abstract :
When a neural network is implemented with limited precision hardware, errors from the quantization of weights become important factors to be considered. In this paper, the authors present several analysis results based on general FNN structures and use several examples to examine the relation between weight errors and output classifications. A lower bound for L, the number of bits used to quantize the weights, is derived in the worst case. This paper also includes the detailed analysis of AND-gates
Keywords :
error analysis; feedforward neural nets; error analysis; feedforward neural networks; lower bound; output classifications; quantized weights; weight errors; worst case; Computer errors; Computer networks; Equations; Error analysis; Feedforward neural networks; Large-scale systems; Neural network hardware; Neural networks; Quantization; Training data;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Southeastcon '94. Creative Technology Transfer - A Global Affair., Proceedings of the 1994 IEEE
Conference_Location :
Miami, FL
Print_ISBN :
0-7803-1797-1
Type :
conf
DOI :
10.1109/SECON.1994.324361
Filename :
324361
Link To Document :
بازگشت