DocumentCode :
296009
Title :
An inverse model learning algorithm using the hierarchical mixtures of experts
Author :
Satoshi, Yamaguchi ; Hidekiyo, Itakura ; Yoshikazu, Nishikawa
Author_Institution :
Dept. of Comput. Sci., Chiba Inst. of Technol., Narashino, Japan
Volume :
5
fYear :
1995
fDate :
Nov/Dec 1995
Firstpage :
2738
Abstract :
A new learning algorithm in neural networks is proposed for inverse modeling. In the learning algorithm, the hierarchical mixtures of experts (HME) is tried out as a forward model of the system. The algorithm is fundamentally based on the back propagation procedure and the updating values of the network synaptic weights are calculated with the help of the HME. Almost all conventional learning algorithms use the Jacobian matrix of the system for estimating the neural network error. This is not the case with our algorithm. As a result, it carefully avoids the local minimum problem which often occurs in some inverse model learning processes
Keywords :
Jacobian matrices; backpropagation; inverse problems; modelling; neural nets; Jacobian matrix; back propagation; hierarchical expert mixtures; inverse model learning algorithm; inverse modeling; network synaptic weights; neural networks; Biological system modeling; Computational biology; Computer science; Inverse problems; Jacobian matrices; Neural networks; Neurofeedback; Supervised learning;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Neural Networks, 1995. Proceedings., IEEE International Conference on
Conference_Location :
Perth, WA
Print_ISBN :
0-7803-2768-3
Type :
conf
DOI :
10.1109/ICNN.1995.488163
Filename :
488163
Link To Document :
بازگشت