DocumentCode :
1694324
Title :
Speaker-independent lips and tongue visualization of vowels
Author :
Hao Li ; Minghao Yang ; Jianhua Tao
Author_Institution :
Nat. Lab. of Pattern Recognition, Inst. of Autom., Beijing, China
fYear :
2013
Firstpage :
8106
Lastpage :
8110
Abstract :
This paper proposes a scheme of speech-driven lips and tongue animation synthesis in a speaker-independent manner. Directional relative displacement (DRD) features are proposed based on the Electromagnetic Articulograph (EMA) data to describe human´s lips and tongue movements, which are more stable across different speakers than the raw EMA data. Multi speakers´ acoustic-articulatory data of vowels are used to learn the acoustic-toarticulatory inversion mapping. We build 2D geometric models of lips and tongue for visualization. With the trained mapping and the geometric models, visualization of lips and tongue movements from acoustic signal of vowels uttered by arbitrary speaker is realized. The experimental results demonstrate that the animations we synthesized are effective aids in helping people identifying vowels.
Keywords :
geometry; learning (artificial intelligence); speech synthesis; 2D geometric model; DRD; EMA data; acoustic-to-articulatory inversion mapping; directional relative displacement; electromagnetic articulograph data; multispeakers acoustic-articulatory data; speaker-independent lips visualization; speaker-independent tongue visualization; speech-driven lips animation synthesis scheme; speech-driven tongue animation synthesis scheme; vowel data; Acoustics; Animation; Coils; Lips; Speech; Tongue; Vectors; acoustic-to-articulatory inversion; articulatory models; electromagnetic articulograph; speech visualization;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Acoustics, Speech and Signal Processing (ICASSP), 2013 IEEE International Conference on
Conference_Location :
Vancouver, BC
ISSN :
1520-6149
Type :
conf
DOI :
10.1109/ICASSP.2013.6639244
Filename :
6639244
Link To Document :
بازگشت