Title :
Online Gesture Spotting from Visual Hull Data
Author :
Peng, Bo ; Qian, Gang
Author_Institution :
Sch. of Arts, Media & Eng., Arizona State Univ., Tempe, AZ, USA
fDate :
6/1/2011 12:00:00 AM
Abstract :
This paper presents a robust framework for online full-body gesture spotting from visual hull data. Using view-invariant pose features as observations, hidden Markov models (HMMs) are trained for gesture spotting from continuous movement data streams. Two major contributions of this paper are 1) view-invariant pose feature extraction from visual hulls, and 2) a systematic approach to automatically detecting and modeling specific nongesture movement patterns and using their HMMs for outlier rejection in gesture spotting. The experimental results have shown the view-invariance property of the proposed pose features for both training poses and new poses unseen in training, as well as the efficacy of using specific nongesture models for outlier rejection. Using the IXMAS gesture data set, the proposed framework has been extensively tested and the gesture spotting results are superior to those reported on the same data set obtained using existing state-of-the-art gesture spotting methods.
Keywords :
data handling; gesture recognition; hidden Markov models; HMM; IXMAS gesture data set; data streams; hidden Markov models; online gesture spotting; visual hull data; Feature extraction; Gesture recognition; Hidden Markov models; Human computer interaction; Tensile stress; Training; Visualization; Online gesture spotting; hidden Markov models; multilinear analysis; nongesture models.; view invariance; visual hull; Algorithms; Artificial Intelligence; Computer Simulation; Gestures; Humans; Image Enhancement; Markov Chains; Models, Biological; Models, Statistical; Movement; Pattern Recognition, Automated;
Journal_Title :
Pattern Analysis and Machine Intelligence, IEEE Transactions on
DOI :
10.1109/TPAMI.2010.199