Title :
Approximated Parallel Model Combination for efficient noise-robust speech recognition
Author_Institution :
Sch. of Comput., Nat. Univ. of Singapore, Singapore, Singapore
Abstract :
Parallel Model Combination (PMC) and Vector Taylor Series (VTS) are two model-based approaches for noise-robust speech recognition. The latter is more popular because of its simple compensation formulae for both the static and dynamic parameters. Furthermore, this VTS compensation formulation can be easily extended to noise adaptive training where the parameters of the underlying pseudo-clean speech and distortion models can be optimized. PMC lacks the above benefits because of its nonlinear variance compensation formula. In this paper, the Approximated PMC (APMC) method is proposed where linearized PMC variance compensation is used. The same approximation has also been applied to Trajectory-based APMC (TAPMC) to achieve a four-time computational saving over the Trajectory-based PMC (TPMC). The dynamic parameter compensation and noise re-estimation formulae for APMC are also derived. Experimental results on AURORA 4 show that APMC and TAPMC consistently outperformed the standard VTS and Trajectory-based VTS (TVTS) by 6.3% and 5.3% relative respectively.
Keywords :
approximation theory; distortion; series (mathematics); speech recognition; AURORA 4; TAPMC; TVTS; VTS compensation; approximated parallel model combination; distortion models; dynamic parameter compensation; linearized PMC variance compensation; noise adaptive training; noise-robust speech recognition; pseudoclean speech; trajectory-based APMC; trajectory-based VTS; vector Taylor series; Approximation methods; Cepstral analysis; Hidden Markov models; Noise; Speech; Speech recognition; Vectors; Noise robust speech recognition; parallel model combination; trajectory-based compensation; vector Taylor series;
Conference_Titel :
Acoustics, Speech and Signal Processing (ICASSP), 2013 IEEE International Conference on
Conference_Location :
Vancouver, BC
DOI :
10.1109/ICASSP.2013.6639097