Title :
Lip movement and speech synchronization detection based on multimodal shift-invariant dictionary
Author :
Zhengyu Zhu; Xiaohui Feng; Jichen Yang
Author_Institution :
South China University of Technology, Guangzhou 510641, China
Abstract :
In order to solve the issue of ignoring the successive and dynamic lip motion information in traditional audio-visual speech synchrony analysis models, a novel method based on shift-invariant learned dictionary is presented. In this method, sparse representation with shift-invariant dictionary is introduced to analyze the bimodal structure of articulation. The learned dictionary is obtained based on the audio-visual coherence dictionary learning algorithm, and the dynamic correlation between voice and lip motion of diverse syllable or word is represented as a pattern by this audio-visual coherence atom. According to these utterance patterns, an original audio-visual synchronization score measuring scheme is proposed. The results of the experiment on four different asynchronous data show the good performance of the method.
Keywords :
"Synchronization","Chlorine","Discrete cosine transforms","Coal"
Conference_Titel :
Communication Technology (ICCT), 2015 IEEE 16th International Conference on
Print_ISBN :
978-1-4673-7004-2
DOI :
10.1109/ICCT.2015.7399944