Title :
Joint language models for automatic speech recognition and understanding
Author :
Bayer, Ali Orkan ; Riccardi, Giuseppe
Author_Institution :
Signals & Interactive Syst. Lab., Univ. of Trento, Trento, Italy
Abstract :
Language models (LMs) are one of the main knowledge sources used by automatic speech recognition (ASR) and Spoken Language Understanding (SLU) systems. In ASR systems they are optimized to decode words from speech for a transcription task. In SLU systems they are optimized to map words into concept constructs or interpretation representations. Performance optimization is generally designed independently for ASR and SLU models in terms of word accuracy and concept accuracy respectively. However, the best word accuracy performance does not always yield the best understanding performance. In this paper we investigate how LMs originally trained to maximize word accuracy can be parametrized to account for speech understanding constraints and maximize concept accuracy. Incremental reduction in concept error rate is observed when a LM is trained on word-to-concept mappings. We show how to optimize the joint transcription and understanding task performance in the lexical-semantic relation space.
Keywords :
decoding; error statistics; natural languages; optimisation; recurrent neural nets; speech recognition; ASR systems; SLU systems; automatic speech recognition; automatic speech understanding; concept accuracy; concept error rate; incremental reduction; interpretation representation; joint language models; joint transcription; knowledge sources; lexical-semantic relation space; performance optimization; transcription task; word accuracy maximization; word decoding; word-to-concept mappings; Accuracy; Joints; Mathematical model; Neural networks; Semantics; Speech; Training; Automatic Speech Recognition; Language Modeling; Recurrent Neural Networks; Spoken Language Understanding;
Conference_Titel :
Spoken Language Technology Workshop (SLT), 2012 IEEE
Conference_Location :
Miami, FL
Print_ISBN :
978-1-4673-5125-6
Electronic_ISBN :
978-1-4673-5124-9
DOI :
10.1109/SLT.2012.6424222