DocumentCode :
2708487
Title :
A gradient-based algorithm competitive with variational Bayesian EM for mixture of Gaussians
Author :
Kuusela, Mikael ; Raiko, Tapani ; Honkela, Antti ; Karhunen, Juha
Author_Institution :
Adaptive Inf. Res. Center, Helsinki Univ. of Technol. (TKK), Helsinki, Finland
fYear :
2009
fDate :
14-19 June 2009
Firstpage :
1688
Lastpage :
1695
Abstract :
While variational Bayesian (VB) inference is typically done with the so called VB EM algorithm, there are models where it cannot be applied because either the E-step or the M-step cannot be solved analytically. In 2007, Honkela et al. introduced a recipe for a gradient-based algorithm for VB inference that does not have such a restriction. In this paper, we derive the algorithm in the case of the mixture of Gaussians model. For the first time, the algorithm is experimentally compared to VB EM and its variant with both artificial and real data. We conclude that the algorithms are approximately as fast depending on the problem.
Keywords :
Gaussian processes; expectation-maximisation algorithm; gradient methods; inference mechanisms; variational techniques; E-step; Gaussian mixture; Gaussian model; M-step; VB EM algorithm; VB inference; gradient based algorithm; gradient-based algorithm; variational Bayesian EM; variational Bayesian inference; Algorithm design and analysis; Bayesian methods; Gaussian processes; Inference algorithms; Least squares approximation; Machine learning; Machine learning algorithms; Neural networks; Probability distribution; Sampling methods;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Neural Networks, 2009. IJCNN 2009. International Joint Conference on
Conference_Location :
Atlanta, GA
ISSN :
1098-7576
Print_ISBN :
978-1-4244-3548-7
Electronic_ISBN :
1098-7576
Type :
conf
DOI :
10.1109/IJCNN.2009.5178726
Filename :
5178726
Link To Document :
بازگشت