DocumentCode :
2196495
Title :
Audio-Visual Speech Synthesis Based on Chinese Visual Triphone
Author :
Zhao, Hui ; Chen, Yue-Bing ; Shen, Ya-Min ; Tang, Chao-Jing
Author_Institution :
Coll. of Electron. Sci. & Eng., Nat. Univ. of Defense Technol., Changsha, China
fYear :
2009
fDate :
17-19 Oct. 2009
Firstpage :
1
Lastpage :
5
Abstract :
A new audio-visual speech synthesis approach is proposed based on Chinese visual triphone. Chinese visual triphone model is constructed using a new clustering method combining artificial immune system and FCM. In the analysis stage, with the training phonetic transcription, visual triphone segments are selected from video sequence, and corresponding lip feature vectors are extracted. In the synthesis stage, viterbi search algorithm is used to select the best visual triphone segments by finding out a path which produces the minimum cost. According to the concatenation principles, mouth animation is generated and stitched into background video. Experimental results show that the synthesized video is natural-looking and satisfactory.
Keywords :
artificial immune systems; audio-visual systems; computer animation; speech synthesis; video signal processing; Chinese visual triphone; FCM; artificial immune system; audio-visual speech synthesis; mouth animation; phonetic transcription; synthesized video; video sequence; viterbi search algorithm; Animation; Artificial immune systems; Clustering algorithms; Clustering methods; Costs; Feature extraction; Mouth; Speech synthesis; Video sequences; Viterbi algorithm;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Image and Signal Processing, 2009. CISP '09. 2nd International Congress on
Conference_Location :
Tianjin
Print_ISBN :
978-1-4244-4129-7
Electronic_ISBN :
978-1-4244-4131-0
Type :
conf
DOI :
10.1109/CISP.2009.5305612
Filename :
5305612
Link To Document :
بازگشت