DocumentCode :
445972
Title :
Finding a succinct multi-layer perceptron having shared weights
Author :
Tanahashi, Yusuke ; Chin, Xiang-Fang ; Saito, Kazumi ; Nakano, Ryohei
Author_Institution :
Nagoya Inst. of Technol., Japan
Volume :
3
fYear :
2005
fDate :
31 July-4 Aug. 2005
Firstpage :
1418
Abstract :
We present a method to find a succinct neural network having shared weights. We focus on weight sharing. Weight sharing constrains the freedom of weight values and weights are allowed to have one of common weights. A near-zero common weight can be eliminated, called weight pruning. Recently, a weight sharing method called BCW has been proposed. The BCW employs merge and split operations based on 2nd-order optimal criteria, and can escape local optima through bidirectional clustering. However, the BCW assumes a vital network parameter J, the number of hidden units, is given. This paper modifies the BCW to make the procedure faster so that the selection of J based on cross-validation can be done in reasonable CPU time. Our experiments showed that the proposed method can restore the original model for an artificial data set, and finds a small number of common weights and an interesting tendency for a real data set.
Keywords :
multilayer perceptrons; bidirectional clustering; multilayer perceptron; weight pruning; weight sharing; Artificial neural networks; Convergence; Data engineering; Data mining; Electronic mail; Laboratories; Multi-layer neural network; Multilayer perceptrons; Neural networks; Polynomials;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Neural Networks, 2005. IJCNN '05. Proceedings. 2005 IEEE International Joint Conference on
Print_ISBN :
0-7803-9048-2
Type :
conf
DOI :
10.1109/IJCNN.2005.1556082
Filename :
1556082
Link To Document :
بازگشت