Title :
Decision level combination of multiple modalities for recognition and analysis of emotional expression
Author :
Metallinou, Angeliki ; Lee, Sungbok ; Narayanan, Shrikanth
Author_Institution :
Dept. of Electr. Eng., Univ. of Southern California, Los Angeles, CA, USA
Abstract :
Emotion is expressed and perceived through multiple modalities. In this work, we model face, voice and head movement cues for emotion recognition and we fuse classifiers using a Bayesian framework. The facial classifier is the best performing followed by the voice and head classifiers and the multiple modalities seem to carry complementary information, especially for happiness. Decision fusion significantly increases the average total unweighted accuracy, from 55% to about 62%. Overall, we achieve average accuracy on the order of 65-75% for emotional states and 30-40% for neutral state using a large multi-speaker, multimodal database. Performance analysis for the case of anger and neutrality suggests a positive correlation between the number of classifiers that performed well and the perceptual salience of the expressed emotion.
Keywords :
Bayes methods; emotion recognition; face recognition; Bayesian framework; decision level combination; emotional expression analysis; emotional expression recognition; multimodal database; multiple modalities; performance analysis; Bayesian methods; Databases; Emotion recognition; Face detection; Fuses; Head; Hidden Markov models; Performance analysis; Robustness; Speech; Bayesian Information Fusion; Hidden Markov Model; Multimodal Emotion Recognition; Perceptual Salience;
Conference_Titel :
Acoustics Speech and Signal Processing (ICASSP), 2010 IEEE International Conference on
Conference_Location :
Dallas, TX
Print_ISBN :
978-1-4244-4295-9
Electronic_ISBN :
1520-6149
DOI :
10.1109/ICASSP.2010.5494890