• DocumentCode
    8192
  • Title

    Unimodal and Multimodal Human Perceptionof Naturalistic Non-Basic Affective Statesduring Human-Computer Interactions

  • Author

    D´Mello, Sidney K. ; Dowell, Nia ; Graesser, Art

  • Author_Institution
    Depts. of Comput. Sci. & Psychol., Univ. of Notre Dame, Notre Dame, IN, USA
  • Volume
    4
  • Issue
    4
  • fYear
    2013
  • fDate
    Oct.-Dec. 2013
  • Firstpage
    452
  • Lastpage
    465
  • Abstract
    The present study investigated unimodal and multimodal emotion perception by humans, with an eye for applying the findings towards automated affect detection. The focus was on assessing the reliability by which untrained human observers could detect naturalistic expressions of non-basic affective states (boredom, engagement/flow, confusion, frustration, and neutral) from previously recorded videos of learners interacting with a computer tutor. The experiment manipulated three modalities to produce seven conditions: face, speech, context, face+speech, face+context, speech+context, face+speech+context. Agreement between two observers (OO) and between an observer and a learner (LO) were computed and analyzed with mixed-effects logistic regression models. The results indicated that agreement was generally low (kappas ranged from .030 to .183), but, with one exception, was greater than chance. Comparisons of overall agreement (across affective states) between the unimodal and multimodal conditions supported redundancy effects between modalities, but there were superadditive, additive, redundant, and inhibitory effects when affective states were individually considered. There was both convergence and divergence of patterns in the OO and LO data sets; however, LO models yielded lower agreement but higher multimodal effects compared to OO models. Implications of the findings for automated affect detection are discussed.
  • Keywords
    behavioural sciences computing; human computer interaction; regression analysis; speech recognition; LO; automated affect detection; computer tutor; face+context; face+speech; face+speech+context; human-computer interactions; learner; mixed-effects logistic regression models; multimodal human perception; naturalistic nonbasic affective states; observer; recorded learners videos; speech+context; unimodal human perception; untrained human observers; Accuracy; Additives; Computers; Context; Face; Speech; Videos; Emotion perception; multimodal affect detection; naturalistic expressions; non-basic emotions; superadditivity;
  • fLanguage
    English
  • Journal_Title
    Affective Computing, IEEE Transactions on
  • Publisher
    ieee
  • ISSN
    1949-3045
  • Type

    jour

  • DOI
    10.1109/T-AFFC.2013.19
  • Filename
    6600687