DocumentCode :
3239450
Title :
Variational expectation-maximization training for Gaussian networks
Author :
Nasios, Nikolaos ; Bors, Adrian G.
Author_Institution :
Dept. of Comput. Sci., York Univ., UK
fYear :
2003
fDate :
17-19 Sept. 2003
Firstpage :
339
Lastpage :
348
Abstract :
This paper introduces variational expectation-maximization (VEM) algorithm for training Gaussian networks. Hyperparameters model distributions of parameters characterizing Gaussian mixture densities. The proposed algorithm employs a hierarchical learning strategy for estimating a set of hyperparameters and the number of Gaussian mixture components. A dual EM algorithm is employed as the initialization stage in the VEM-based learning. In the first stage the EM algorithm is applied on the given data set while the second stage EM is used on distributions of parameters resulted from several runs of the first stage EM. Appropriate maximum log-likelihood estimators are considered for all the parameter distributions involved.
Keywords :
Gaussian distribution; learning (artificial intelligence); maximum likelihood estimation; optimisation; Gaussian mixture densities; Gaussian networks; hierarchical learning strategy; hyperparameters model distributions; maximum log-likelihood estimation; variational expectation-maximization algorithm; Approximation algorithms; Bayesian methods; Computer science; Distributed computing; Kernel; Parameter estimation; Probability distribution; Radial basis function networks; Stochastic processes; Uncertainty;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Neural Networks for Signal Processing, 2003. NNSP'03. 2003 IEEE 13th Workshop on
ISSN :
1089-3555
Print_ISBN :
0-7803-8177-7
Type :
conf
DOI :
10.1109/NNSP.2003.1318033
Filename :
1318033
Link To Document :
بازگشت