Title :
Revisiting graphemes with increasing amounts of data
Author :
Sung, Yun-hsuan ; Hughes, Thad ; Beaufays, Françoise ; Strope, Brian
Author_Institution :
Dept. of EE, Stanford Univ., Stanford, CA
Abstract :
Letter units, or graphemes, have been reported in the literature as a surprisingly effective substitute to the more traditional phoneme units, at least in languages that enjoy a strong correspondence between pronunciation and orthography. For English however, where letter symbols have less acoustic consistency, previously reported results fell short of systems using highly-tuned pronunciation lexicons. Grapheme units simplify system design, but since graphemes map to a wider set of acoustic realizations than phonemes, we should expect grapheme-based acoustic models to require more training data to capture these variations. In this paper, we compare the rate of improvement of grapheme and phoneme systems trained with datasets ranging from 450 to 1200 hours of speech. We consider various grapheme unit configurations, including using letter-specific, onset, and coda units. We show that the grapheme systems improve faster and, depending on the lexicon, reach or surpass the phoneme baselines with the largest training set.
Keywords :
speech recognition; acoustic consistency; coda units; grapheme system; grapheme unit configuration; grapheme-based acoustic model; highly-tuned pronunciation lexicons; letter symbols; letter units; letter-specific units; onset units; orthography; phoneme units; speech recognition; Acoustics; Context modeling; Costs; Gaussian processes; Natural languages; Probability; Scalability; Speech recognition; Training data; Web search; Acoustic modeling; directory assistance; graphemes; speech recognition;
Conference_Titel :
Acoustics, Speech and Signal Processing, 2009. ICASSP 2009. IEEE International Conference on
Conference_Location :
Taipei
Print_ISBN :
978-1-4244-2353-8
Electronic_ISBN :
1520-6149
DOI :
10.1109/ICASSP.2009.4960617