DocumentCode :
329055
Title :
Explicit solution of the optimum weights of multilayer perceptron: the binary input case
Author :
Yu, Xiao-Hu ; Guo, Qiang
Author_Institution :
Nat. Commun. Res. Lab., Southeast Univ., Nanjing, China
Volume :
2
fYear :
1993
fDate :
25-29 Oct. 1993
Firstpage :
1689
Abstract :
Explicit solutions of multilayer feedforward networks have previously been discussed by Yu (1992). This paper extends the similar idea to binary input case. We show that the hidden units can be used to form the basis functions of the binary Walsh transform and the network training can therefore be treated as finding the coefficients of the binary Walsh expansion of the desired mapping, thus making the optimum weights explicitly solvable. For the incomplete training set case, a useful approach is presented to assure the resultant network having smooth generalization performance. Noise rejection performance of the obtained network is also illustrated.
Keywords :
Walsh functions; feedforward neural nets; generalisation (artificial intelligence); learning (artificial intelligence); multilayer perceptrons; transforms; binary Walsh transform; binary input; generalization; hidden units; mapping; multilayer feedforward networks; multilayer perceptron; network learning; noise rejection; optimum weights; Backpropagation algorithms; Computer aided software engineering; Equations; Multilayer perceptrons; Nonhomogeneous media; Radar; Sufficient conditions;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Neural Networks, 1993. IJCNN '93-Nagoya. Proceedings of 1993 International Joint Conference on
Print_ISBN :
0-7803-1421-2
Type :
conf
DOI :
10.1109/IJCNN.1993.716978
Filename :
716978
Link To Document :
بازگشت