DocumentCode
394369
Title
Covariance and precision modeling in shared multiple subspaces
Author
Dharanipragada, S. ; Visweswariah, K.
Author_Institution
IBM T. J. Watson Res. Center, Yorktown Heights, NY, USA
Volume
1
fYear
2003
fDate
6-10 April 2003
Abstract
We introduce a class of Gaussian mixture models for HMM states in continuous speech recognition. In these models, the covariances or the precisions (inverse covariances) are restricted to lie in subspaces spanned by rank-one symmetric matrices. In both cases, the rank-one matrices are shared across classes of Gaussians. We show that, for the same number of parameters, modeling precisions leads to better performance when compared to modeling covariances. Modeling precisions however gives a distinct advantage in computational and memory requirements. We also show that this class of models provides improvement in accuracy (for the same number of parameters) over classical factor analysed models and the recently proposed EMLLT (extended maximum likelihood linear transform) models which are special instances of this class of models.
Keywords
Gaussian processes; covariance matrices; hidden Markov models; speech recognition; Gaussian mixture models; HMM; computational requirements; continuous speech recognition; covariance modeling; extended maximum likelihood linear transform models; memory requirements; precision modeling; rank-one symmetric matrices; shared multiple subspaces; Closed-form solution; Computational complexity; Covariance matrix; Hidden Markov models; Maximum likelihood estimation; Parameter estimation; Robustness; Speech recognition; Symmetric matrices; Unsolicited electronic mail;
fLanguage
English
Publisher
ieee
Conference_Titel
Acoustics, Speech, and Signal Processing, 2003. Proceedings. (ICASSP '03). 2003 IEEE International Conference on
ISSN
1520-6149
Print_ISBN
0-7803-7663-3
Type
conf
DOI
10.1109/ICASSP.2003.1198916
Filename
1198916
Link To Document