Title :
Training MLPs via the expectation maximization algorithm
Author :
Cook, G.D. ; Robinson, A.J.
Author_Institution :
Cambridge Univ., UK
Abstract :
Two factors that inhibit the use of multi layer perceptrons (MLP) by the pattern recognition and statistical modelling community are the long training times and a poorly developed theory of network internals. The paper proposes a solution to these problems by introducing a framework in which the hidden nodes of a multi layer perceptron are models of a binary stochastic process governed by the Bernoulli probability density function. Within this framework, we present an expectation maximisation (EM) algorithm for training the network. The algorithm iteratively finds the most probable internal representation given the network inputs and desired outputs. The internal representation forms target values for the hidden nodes so that the desired output of all nodes in the network is specified. This removes the necessity of using back propagation of errors to modify hidden layer weights, and effectively decomposes the MLP into two single layer perceptrons, the parameters of which can be efficiently optimised. We demonstrate the use of our algorithm on a number of binary and multiway classification tasks
Keywords :
iterative methods; learning (artificial intelligence); multilayer perceptrons; probability; stochastic processes; Bernoulli probability density function; MLP training; binary stochastic process; desired outputs; expectation maximization algorithm; hidden nodes; multi layer perceptrons; multiway classification tasks; network inputs; network internals; pattern recognition; probable internal representation; single layer perceptrons; statistical modelling;
Conference_Titel :
Artificial Neural Networks, 1995., Fourth International Conference on
Conference_Location :
Cambridge
Print_ISBN :
0-85296-641-5
DOI :
10.1049/cp:19950527