DocumentCode :
518625
Title :
Speech-to-visual speech synthesis using Chinese visual triphone
Author :
Zhao, Hui ; Shen, Yamin ; Tang, Chaojing
Author_Institution :
Coll. of Electron. Sci. & Eng., Nat. Univ. of Defense Technol., Changsha, China
Volume :
2
fYear :
2010
fDate :
27-29 March 2010
Firstpage :
241
Lastpage :
245
Abstract :
A visual speech synthesis approach using Chinese visual triphone is presented. According to Mandarin Chinese pronunciation principle and the relationship between phoneme and viseme, “Chinese visual triphone” model is constructed. Triphone hidden Markov model is established based on visual triphones. Joint features composed of visual features and audio features are used in the training stage. In the synthesis stage, sentence HMM is constructed by concatenating triphone HMMs. With the features extracted from sentence HMM, visual speech is synthesized. From the scores of subjective and objective estimation, the synthesized video is realistic and satisfactory.
Keywords :
hidden Markov models; speech synthesis; Chinese visual triphone; Mandarin Chinese pronunciation principle; hidden Markov model; objective estimation; speech-to-visual speech synthesis; synthesized video; Acoustic noise; Auditory system; Chaos; Educational institutions; Feature extraction; Hidden Markov models; Natural languages; Signal synthesis; Speech synthesis; Working environment noise; Chinese visual triphone; hidden Markov model (HMM); joint features; visual speech synthesis;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Advanced Computer Control (ICACC), 2010 2nd International Conference on
Conference_Location :
Shenyang
Print_ISBN :
978-1-4244-5845-5
Type :
conf
DOI :
10.1109/ICACC.2010.5486681
Filename :
5486681
Link To Document :
بازگشت