DocumentCode
43017
Title
Automatic Detection of Nonverbal Behavior Predicts Learning in Dyadic Interactions
Author
Won, Andrea Stevenson ; Bailenson, Jeremy N. ; Janssen, Joris H.
Author_Institution
Dept. of Commun., Stanford Univ., Stanford, CA, USA
Volume
5
Issue
2
fYear
2014
fDate
April-June 1 2014
Firstpage
112
Lastpage
125
Abstract
Nonverbal behavior can reveal the psychological states of those engaged in interpersonal interaction. Previous research has highlighted the relationship between gesture and learning during instruction. In the current study we applied readily available computer vision hardware and machine learning algorithms to the gestures of teacher/student dyads (N = 106) during a learning session to automatically distinguish between high and low success learning interactions, operationalized by recall for information presented during that learning session. Models predicted learning performance of the dyad with accuracies as high as 85.7 percent when tested on dyads not included in the training set. In addition, correlations were found between summed measures of body movement and learning score. We discuss theoretical and applied implications for learning.
Keywords
computer vision; gesture recognition; human computer interaction; learning (artificial intelligence); object detection; psychology; automatic detection; body movement; computer vision hardware; dyadic interactions; gesture recognition; high success learning interactions; interpersonal interaction; learning performance; learning score; learning session; low success learning interactions; machine learning algorithms; nonverbal behavior; psychological states; Computer vision; Education; Feature extraction; Joints; Materials; Observers; Tracking; Natural data set; collaborative learning; gesture recognition; machine learning;
fLanguage
English
Journal_Title
Affective Computing, IEEE Transactions on
Publisher
ieee
ISSN
1949-3045
Type
jour
DOI
10.1109/TAFFC.2014.2329304
Filename
6827904
Link To Document