Title :
Network expansion and network compression-further discussion on structure variation methods
Author_Institution :
Inst. of Comput. Sci. & Technol., Peking Univ., Beijing, China
Abstract :
There has been an increasing tendency of varying the structure of networks during training processes. The author has divided the structure variation methods in feedforward neural networks into three parts-network expansion, network construction, and network compression and presents some methods. Since there are numerous papers concerning network construction, in this paper, the author only further addresses two parts: network expansion and network compression. The author groups network expansion methods into two classes: retraining all the weights and only training the added weights after the network is expanded. For network compression, the author first divides the compression process into two stages: in the first stage the author selects which connection or neuron should be moved. In this stage, the compression strategies are classified into three groups. In the second stage, the author prunes the selected one. In this stage, the author proceeds by another strategy in which the pruning methods are classified into two groups: pruning without information crosswise propagation (CP) and pruning with information CP. Combination of network expansion and network compression is generally used in practice. As a special case of combination of network expansion and network compression, a duplication problem is addressed
Keywords :
feedforward neural nets; genetic algorithms; learning (artificial intelligence); duplication problem; feedforward neural networks; information crosswise propagation; network compression; network expansion; pruning; structure variation methods; training processes; Computer science; Feedforward neural networks; Linearity; Multilayer perceptrons; Neural networks; Neurons;
Conference_Titel :
Neural Networks, 1995. Proceedings., IEEE International Conference on
Conference_Location :
Perth, WA
Print_ISBN :
0-7803-2768-3
DOI :
10.1109/ICNN.1995.488262