Title :
A novel scheme to determine the architecture of a multilayer perceptron
Author :
Chintalapudi, K.K. ; Pal, N.R.
Author_Institution :
Dept. of Electr. & Comput. Eng., Drexel Univ., Philadelphia, PA, USA
Abstract :
We propose a method for optimizing the architecture of a multilayer perceptron (MLP) network. The proposed scheme is a variation of the MLP architecture, in which each neuron´s output is modulated by an efficiency factor associated with that node. Nodes with low efficiency factor literally do not participate in the network. We compute the efficiency of a node using a multiplier function with a learnable parameter, which we call the multiplier of that node. Values of the multipliers are learned by a gradient descent along with the weights, aiming to minimize the mean square error. Training starts with all node efficiencies set very low so that there is literally no connection between any of the neurons in the net. As the learning progresses, gradually some of the nodes start acquiring high efficiencies. Training is terminated when performance of the network is satisfactory. At the end of the training nodes with low efficiency are eliminated and a near optimal architectural size for the MLP is obtained. Effectiveness of the proposed scheme is demonstrated on several data-sets
Keywords :
gradient methods; learning (artificial intelligence); multilayer perceptrons; neural net architecture; optimisation; efficiency factor; gradient descent method; learning process; mean square error; multilayer perceptron; multiplier function; neural net architecture; optimisation; Computer architecture; Machine intelligence; Mean square error methods; Multilayer perceptrons; Neurons; Optimization methods; Training data; Weight measurement;
Conference_Titel :
Systems, Man, and Cybernetics, 1998. 1998 IEEE International Conference on
Conference_Location :
San Diego, CA
Print_ISBN :
0-7803-4778-1
DOI :
10.1109/ICSMC.1998.724998