Title :
Psychoacoustic model compensation with robust feature set for speaker verification in additive noise
Author_Institution :
TCS Innovation Lab., Mumbai, India
Abstract :
This paper addresses the problem of speaker verification in the presence of additive noise for resource deficient languages. Psychoacoustic model compensation (Psy-Comp) has been shown to impart noise robustness to Gaussian Mixture Model (GMM) based speaker verification systems using Mel Frequency Cepstral Coefficients (MFCCs). This work extends the idea of Psy-Comp to incorporate a more robust feature set, which includes Cepstral Mean Subtraction (CMS) and Δ coefficients along with the MFCCs. We propose a model domain CMS operation following the psychoacoustic compensation for improved performance in additive noise. An advantage of this approach is that it does not require specialized developmental data and hence it may be suitable for resource deficient languages. Experiments conducted with the NIST-2000 database corrupted with real-life street noise show improved performance with the proposed method.
Keywords :
Gaussian processes; acoustic signal processing; cepstral analysis; mixture models; natural language processing; psychology; speaker recognition; CMS; GMM based speaker verification systems; Gaussian mixture model; MFCC; NIST-2000 database; Psy-Comp; additive noise; cepstral mean subtraction; mel frequency cepstral coefficients; noise robustness; psychoacoustic model compensation; real-life street noise; resource deficient languages; robust feature set; Computational modeling; Mathematical model; Noise; Psychoacoustic models; Speech; Training; Vectors; Additive Noise; Psychoacoustics; Speaker Recognition;
Conference_Titel :
Chinese Spoken Language Processing (ISCSLP), 2014 9th International Symposium on
Conference_Location :
Singapore
DOI :
10.1109/ISCSLP.2014.6936706