DocumentCode
296063
Title
Hierarchical mixture of experts and Max-Min propagation neural networks
Author
Estévez, Pablo A. ; Nakano, Ryohei
Author_Institution
Dept. of Electr. Eng., Chile Univ., Santiago, Chile
Volume
1
fYear
1995
fDate
Nov/Dec 1995
Firstpage
651
Abstract
The max-min propagation neural network model is considered as a hierarchical mixture of experts by replacing the max (min) units with softmax functions. The resulting mixture is different from the model of Jordan and Jacobs, but we exploit the similarities between both models to derive a probability model. Learning is treated as a maximum-likelihood problem, in particular we present a gradient ascent algorithm and an expectation-maximization algorithm. Simulation results on the parity problem and the majority problem are reported
Keywords
learning (artificial intelligence); maximum likelihood estimation; minimax techniques; neural nets; probability; expectation-maximization algorithm; gradient ascent algorithm; hierarchical mixture of experts; learning algorithm; majority problem; max-min propagation neural network; maximum-likelihood; parity problem; probability model; softmax function; Electronic mail; Expectation-maximization algorithms; Jacobian matrices; Laboratories; Least squares approximation; Least squares methods; Multilayer perceptrons; Neural networks; Vectors;
fLanguage
English
Publisher
ieee
Conference_Titel
Neural Networks, 1995. Proceedings., IEEE International Conference on
Conference_Location
Perth, WA
Print_ISBN
0-7803-2768-3
Type
conf
DOI
10.1109/ICNN.1995.488257
Filename
488257
Link To Document