Title :
Dynamic process modelling using locally recurrent neural networks
Author :
Zhang, Jie ; Morris, A. Julian
Author_Institution :
Dept. of Chem. & Process Eng., Newcastle upon Tyne Univ., UK
Abstract :
A sequential orthogonal training method is developed in this paper to train locally recurrent neural networks which contain a single hidden layer and dynamic neurons are located in the hidden layer. During network training, the first hidden neuron is used to model the relationship between inputs and outputs whereas other hidden neurons are added sequentially to model the relationship between inputs and model residues. When adding a hidden neuron, its contribution is due to that part of its output vector which is orthogonal to the space spanned by the output vectors of the previous hidden neurons. The Gram-Schmidt orthogonalisation technique is used at each training step to form a set of orthogonal bases for the space spanned by the hidden neuron outputs. The optimum hidden layer weights can be obtained through a gradient based optimisation method while the output layer weights can be found using least squares regression. Hidden neurons are added sequentially and the training procedure terminates when the model error is lower than a pre-set level. Using this training method, neurons with mixed types of activation functions and dynamic orders can be incorporated into a single network. Mixed node networks can offer improved performance in terms of representation capabilities and network size. The excellent performance of the proposed technique is demonstrated by application examples
Keywords :
learning (artificial intelligence); modelling; optimisation; prediction theory; recurrent neural nets; transfer functions; Gram-Schmidt orthogonalisation technique; dynamic orders; dynamic process modelling; first hidden neuron; gradient based optimisation method; least squares regression; locally recurrent neural networks; mixed activation functions; mixed node networks; network size; network training; optimum hidden layer weights; representation capabilities; sequential orthogonal training method; single hidden layer; Chemical engineering; Chemical processes; Electronic mail; Feedforward neural networks; Feeds; Neural networks; Neurons; Optimization methods; Predictive models; Recurrent neural networks;
Conference_Titel :
American Control Conference, Proceedings of the 1995
Conference_Location :
Seattle, WA
Print_ISBN :
0-7803-2445-5
DOI :
10.1109/ACC.1995.532353