DocumentCode
730797
Title
Leveraging automatic speech recognition in cochlear implants for improved speech intelligibility under reverberation
Author
Hazrati, Oldooz ; Ghaffarzadegan, Shabnam ; Hansen, John H. L.
Author_Institution
Center for Robust Speech Syst. (CRSS), Univ. of Texas at Dallas, Richardson, TX, USA
fYear
2015
fDate
19-24 April 2015
Firstpage
5093
Lastpage
5097
Abstract
Despite recent advancements in digital signal processing technology for cochlear implant (CI) devices, there still remains a significant gap between speech identification performance of CI users in reverberation compared to that in anechoic quiet conditions. Alternatively, automatic speech recognition (ASR) systems have seen significant improvements in recent years resulting in robust speech recognition in a variety of adverse environments, including reverberation. In this study, we exploit advancements seen in ASR technology for alternative formulated solutions to benefit CI users. Specifically, an ASR system is developed using multicondition training on speech data with different reverberation characteristics (e.g., T60 values), resulting in low word error rates (WER) in reverberant conditions. A speech synthesizer is then utilized to generate speech waveforms from the output of the ASR system, from which the synthesized speech is presented to CI listeners. The effectiveness of this hybrid recognition-synthesis CI strategy is evaluated under moderate to highly reverberant conditions (i.e., T60 = 0.3, 0.6, 0.8, and 1.0s) using speech material extracted from the TIMIT corpus. Experimental results confirm the effectiveness of multi-condition training on performance of the ASR system in reverberation, which consequently results in substantial speech intelligibility gains for CI users in reverberant environments.
Keywords
acoustic noise; cochlear implants; medical signal processing; reverberation; speech intelligibility; speech processing; speech recognition; speech synthesis; ASR system output; ASR technology; TIMIT corpus; automatic speech recognition; cochlear implants; digital signal processing technology; hybrid recognition-synthesis strategy; improved speech intelligibility; multicondition training; reverberation conditions; speech data; speech identification performance; speech synthesizer; speech waveform generation; word error rate; Cochlear implants; Noise; Reverberation; Speech; Speech enhancement; Speech recognition; Training; Automatic speech recognition; cochlear implants; multi-condition training; reverberation;
fLanguage
English
Publisher
ieee
Conference_Titel
Acoustics, Speech and Signal Processing (ICASSP), 2015 IEEE International Conference on
Conference_Location
South Brisbane, QLD
Type
conf
DOI
10.1109/ICASSP.2015.7178941
Filename
7178941
Link To Document