DocumentCode :
3703319
Title :
EmoShapelets: Capturing local dynamics of audio-visual affective speech
Author :
Yuan Shangguan;Emily Mower Provost
Author_Institution :
University of Michigan, Ann Arbor, Michigan 48109
fYear :
2015
Firstpage :
229
Lastpage :
235
Abstract :
Automatic recognition of emotion in speech is an active area of research. One of the important open challenges relates to how the emotional characteristics of speech change in time. Past research has demonstrated the importance of capturing global dynamics (across an entire utterance) and local dynamics (within segments of an utterance). In this paper, we propose a novel concept, EmoShapelets, to capture the local dynamics in speech. EmoShapelets capture changes in emotion that occur within utterances. We propose a framework to generate, update, and select EmoShapelets. We also demonstrate the discriminative power of EmoShapelets by using them with various classifiers to achieve comparable results with the state-of-the-art systems on the IEMOCAP dataset. EmoShapelets can serve as basic units of emotion expression and provide additional evidence supporting the existence of local patterns of emotion underlying human communication.
Keywords :
"Time series analysis","Speech","Hidden Markov models","Feature extraction","Speech recognition","Heuristic algorithms","Training"
Publisher :
ieee
Conference_Titel :
Affective Computing and Intelligent Interaction (ACII), 2015 International Conference on
Electronic_ISBN :
2156-8111
Type :
conf
DOI :
10.1109/ACII.2015.7344576
Filename :
7344576
Link To Document :
بازگشت