Title : 
Variable n-grams and extensions for conversational speech language modeling
         
        
            Author : 
Siu, Manhung ; Ostendorf, Mari
         
        
            Author_Institution : 
Hong Kong Univ. of Sci. & Technol., Kowloon, Hong Kong
         
        
        
        
        
            fDate : 
1/1/2000 12:00:00 AM
         
        
        
        
            Abstract : 
Recent progress in variable n-gram language modeling provides an efficient representation of n-gram models and makes training of higher order n grams possible. We apply the variable n-gram design algorithm to conversational speech, extending the algorithm to learn skips and context-dependent classes to handle conversational speech characteristics such as filler words, repetitions, and other disfluencies. Experiments show that using the extended variable n-gram results in a language model that captures 4-gram context with less than half the parameters of a standard trigram while also improving the test perplexity and recognition accuracy
         
        
            Keywords : 
computational linguistics; natural languages; speech recognition; context-dependent classes; conversational speech language modeling; experiments; speech recognition; training; trigram; variable n-gram language modeling; Algorithm design and analysis; Context modeling; Costs; Helium; History; Natural languages; Parameter estimation; Speech recognition; Testing; Vocabulary;
         
        
        
            Journal_Title : 
Speech and Audio Processing, IEEE Transactions on