DocumentCode :
231557
Title :
Improvements on bottleneck feature for large vocabulary continuous speech recognition
Author :
Tuerxun, Maimaitiaili ; Shiliang Zhang ; Yebo Bao ; Lirong Dai
Author_Institution :
Nat. Eng. Lab. for Speech & Language Inf. Process., Univ. of Sci. & Technol. of China, Hefei, China
fYear :
2014
fDate :
19-23 Oct. 2014
Firstpage :
516
Lastpage :
520
Abstract :
In this paper, we have proposed three methods to improve the performance of the bottleneck(BN) feature based GMM-HMM system. Firstly, we recommend a new bottleneck feature architecture, namely LBN, which places the bottleneck layer at the last hidden layer instead of the middle, in order to take advantage of the more discriminative and invariant higher layer features. Secondly, we employ the rectified linear units (ReLUs) based DNN as bottleneck feature extractor. Finally, we investigate the sequence discriminative training of bottleneck neural network to achieve more powerful bottleneck feature. We have evaluated our methods in 309-hour Switchboard (SWB) task. Compared with the traditional hybrid DNN-HMM system, our proposed ReLUs based LBN-GMM-HMM system can achieve about 9.7% recognition error rate reduction relatively.
Keywords :
hidden Markov models; neural nets; speech recognition; 309-hour switchboard task; ReLUs based LBN-GMM-HMM system; bottleneck feature based GMM-HMM system; bottleneck neural network; large vocabulary continuous speech recognition; last hidden layer; rectified linear units; Acoustics; Computer architecture; Feature extraction; Hidden Markov models; Neural networks; Speech recognition; Training; DNN; LBN-GMM-HMM; bottleneck feature (BN); rectified linear units (ReLUs); sequence discriminative training;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Signal Processing (ICSP), 2014 12th International Conference on
Conference_Location :
Hangzhou
ISSN :
2164-5221
Print_ISBN :
978-1-4799-2188-1
Type :
conf
DOI :
10.1109/ICOSP.2014.7015058
Filename :
7015058
Link To Document :
بازگشت