DocumentCode
1589944
Title
Performance Improvement in Speech Recognition Using Multimodal Features
Author
Kim, Myung Won ; Song, Won Moon ; Kim, Young Jin ; Kim, Eun Ju
Author_Institution
Soongsil Univ., Seoul
Volume
2
fYear
2007
Firstpage
686
Lastpage
690
Abstract
In this paper, we propose a neural network based model of robust speech recognition by integrating audio, visual, and contextual information. bimodal neural network(BMNN) is a multi-layer perceptron of 4 layers, which combines audio and visual features of speech to compensate loss of audio information caused by noise. In order to improve the accuracy of speech recognition in noisy environments, we also propose a post-processing based on contextual information which are sequential patterns of words spoken by a user. Our experimental results show that our model outperforms any single mode models. Particularly, when we use the contextual information, we can obtain over 90% recognition accuracy even in noisy environments, which is a significant improvement compared with the state of art in speech recognition.
Keywords
feature extraction; multilayer perceptrons; speech processing; speech recognition; audio features; bimodal neural network; contextual information; multi-layer perceptron; multimodal features; neural network based model; noisy environments; sequential patterns; speech recognition; visual features; Computer networks; Context modeling; Electronic mail; Fuses; Hidden Markov models; Moon; Neural networks; Noise robustness; Speech recognition; Working environment noise;
fLanguage
English
Publisher
ieee
Conference_Titel
Natural Computation, 2007. ICNC 2007. Third International Conference on
Conference_Location
Haikou
Print_ISBN
978-0-7695-2875-5
Type
conf
DOI
10.1109/ICNC.2007.550
Filename
4344438
Link To Document