DocumentCode
8192
Title
Unimodal and Multimodal Human Perceptionof Naturalistic Non-Basic Affective Statesduring Human-Computer Interactions
Author
D´Mello, Sidney K. ; Dowell, Nia ; Graesser, Art
Author_Institution
Depts. of Comput. Sci. & Psychol., Univ. of Notre Dame, Notre Dame, IN, USA
Volume
4
Issue
4
fYear
2013
fDate
Oct.-Dec. 2013
Firstpage
452
Lastpage
465
Abstract
The present study investigated unimodal and multimodal emotion perception by humans, with an eye for applying the findings towards automated affect detection. The focus was on assessing the reliability by which untrained human observers could detect naturalistic expressions of non-basic affective states (boredom, engagement/flow, confusion, frustration, and neutral) from previously recorded videos of learners interacting with a computer tutor. The experiment manipulated three modalities to produce seven conditions: face, speech, context, face+speech, face+context, speech+context, face+speech+context. Agreement between two observers (OO) and between an observer and a learner (LO) were computed and analyzed with mixed-effects logistic regression models. The results indicated that agreement was generally low (kappas ranged from .030 to .183), but, with one exception, was greater than chance. Comparisons of overall agreement (across affective states) between the unimodal and multimodal conditions supported redundancy effects between modalities, but there were superadditive, additive, redundant, and inhibitory effects when affective states were individually considered. There was both convergence and divergence of patterns in the OO and LO data sets; however, LO models yielded lower agreement but higher multimodal effects compared to OO models. Implications of the findings for automated affect detection are discussed.
Keywords
behavioural sciences computing; human computer interaction; regression analysis; speech recognition; LO; automated affect detection; computer tutor; face+context; face+speech; face+speech+context; human-computer interactions; learner; mixed-effects logistic regression models; multimodal human perception; naturalistic nonbasic affective states; observer; recorded learners videos; speech+context; unimodal human perception; untrained human observers; Accuracy; Additives; Computers; Context; Face; Speech; Videos; Emotion perception; multimodal affect detection; naturalistic expressions; non-basic emotions; superadditivity;
fLanguage
English
Journal_Title
Affective Computing, IEEE Transactions on
Publisher
ieee
ISSN
1949-3045
Type
jour
DOI
10.1109/T-AFFC.2013.19
Filename
6600687
Link To Document