Title : 
Integrating detailed information into a language model
         
        
            Author : 
Zhang, Ruiqiang ; Black, Ezra ; Finch, Andrew ; Sagisaka, Yoshinori
         
        
            Author_Institution : 
ATR Interpreting Telephony Res. Labs., Kyoto, Japan
         
        
        
        
        
        
            Abstract : 
Applying natural language processing technique to language modeling is a key problem in speech recognition. This paper describes a maximum entropy-based approach to language modeling in which both words together with syntactic and semantic tags in the long history are used as a basis for complex linguistic questions. These questions are integrated with a standard trigram language model or a standard trigram language model combined with long history word triggers and the resulting language model is used to rescore the N-best hypotheses output of the ATRSPREC speech recognition system. The technique removed 24% of the correctable error of the recognition system
         
        
            Keywords : 
linguistics; maximum entropy methods; natural languages; speech recognition; ATRSPREC speech recognition system; N-best hypotheses output; complex linguistic questions; correctable error; language model; long history word triggers; maximum entropy-based approach; natural language processing technique; semantic tags; speech recognition; standard trigram language model; syntactic tags; Error analysis; Error correction; History; Information resources; Laboratories; Natural language processing; Natural languages; Predictive models; Speech recognition; Vocabulary;
         
        
        
        
            Conference_Titel : 
Acoustics, Speech, and Signal Processing, 2000. ICASSP '00. Proceedings. 2000 IEEE International Conference on
         
        
            Conference_Location : 
Istanbul
         
        
        
            Print_ISBN : 
0-7803-6293-4
         
        
        
            DOI : 
10.1109/ICASSP.2000.861995