DocumentCode
2748949
Title
Analysis of the effects of quantization in high-order function neural network
Author
Minghu, Jiang ; Xiaoyan, Zhu ; Ying, Lin ; Baozong, Yuan ; Xiaofang, Tang ; Biqin, Lin ; Qiuqi, Ruan ; Mingyan, Jiang
Author_Institution
Inst. of Inf. Sci., Northern Jiaotong Univ., Beijing, China
Volume
3
fYear
2000
fDate
2000
Firstpage
1629
Abstract
A statistical quantization model is used to analyze of the effects of quantization in digital implementation of high-order function neural network. From the theory we analyse the performance degradation and fault tolerance of the neural network caused by the number of quantization bits and by changing the order. We try to predict the error in the high-order function neural network (HOFNN) given the properties of the network and the number of quantization bits. Experimental results show the error rate is inversely proportional to quantized bits M for HRFNN. The recognition performance of the backpropagation (BP) network and the HRFNN are almost the same for different quantization bits. The network´s performance degradation gets worse when the number of bits is lower than 4-bit quantization. The network´s performance degradation gets worse when the number of bits is lower than 4-bit quantization
Keywords
backpropagation; neural nets; quantisation (signal); statistical analysis; HOFNN; HRFNN; backpropagation network; digital implementation; error rate; fault tolerance; high-order function neural network; performance degradation; quantization bits; recognition performance; statistical quantization model; Computer networks; Degradation; Fault tolerance; Information analysis; Intelligent networks; Intelligent systems; Neural networks; Neurons; Performance analysis; Quantization;
fLanguage
English
Publisher
ieee
Conference_Titel
Signal Processing Proceedings, 2000. WCCC-ICSP 2000. 5th International Conference on
Conference_Location
Beijing
Print_ISBN
0-7803-5747-7
Type
conf
DOI
10.1109/ICOSP.2000.893413
Filename
893413
Link To Document