Title :
Using deformable templates to infer visual speech dynamics
Author :
Hennecke, Marcus E. ; Prasad, K. Venkatesh ; Stork, David G.
Author_Institution :
Dept. of Electr. Eng., Stanford Univ., CA, USA
fDate :
31 Oct-2 Nov 1994
Abstract :
The visual image of a talker provides information complementary to the acoustic speech waveform, and enables improved recognition accuracy, especially in environments corrupted by high acoustic noise or multiple talkers. Because most of the phonologically relevant visual information is from the mouth and lips, it is important to infer accurately and robustly their dynamics; moreover it is desirable to extract this information without the use of invasive markers or patterned illumination. We describe the use of deformable templates for speechreading, in order to infer the dynamics of lip contours throughout an image sequence. Template computations can be done relatively quickly and the resulting small number of shape description parameters are quite robust to visual noise and variations in illumination. Such templates delineate the inside of the mouth, so that the teeth and the tongue can also be found
Keywords :
acoustic noise; acoustic signal processing; dynamics; image sequences; speech processing; speech recognition; vision; acoustic speech waveform; automatic speech recognition; cost function; deformable templates; high acoustic noise; illumination variations; image sequence; lip contours; mouth; multiple talkers; phonologically relevant visual information; recognition accuracy; shape description parameters; speechreading; teeth; template computations; tongue; visual image; visual noise; visual speech dynamics; Acoustic noise; Acoustic waves; Data mining; Image recognition; Lighting; Lips; Mouth; Noise robustness; Speech enhancement; Speech recognition;
Conference_Titel :
Signals, Systems and Computers, 1994. 1994 Conference Record of the Twenty-Eighth Asilomar Conference on
Conference_Location :
Pacific Grove, CA
Print_ISBN :
0-8186-6405-3
DOI :
10.1109/ACSSC.1994.471518