• DocumentCode
    2240753
  • Title

    Space-time gestures

  • Author

    Darrell, Trevor ; Pentland, Alex

  • Author_Institution
    MIT Media Lab., MA, USA
  • fYear
    1993
  • fDate
    15-17 Jun 1993
  • Firstpage
    335
  • Lastpage
    340
  • Abstract
    A method for learning, tracking, and recognizing human gestures using a view-based approach to model articulated objects is presented. Objects are represented using sets of view models, rather than single templates. Stereotypical space-time patterns, i.e., gestures, are then matched to stored gesture patterns using dynamic time warping. Real-time performance is achieved by using special purpose correlation hardware and view prediction to prune as much of the search space as possible. Both view models and view predictions are learned from examples. Results showing tracking and recognition of human hand gestures at over 10 Hz are presented
  • Keywords
    correlators; human factors; image recognition; image sequences; motion estimation; real-time systems; user interfaces; 10 Hz; articulated objects; correlation hardware; dynamic time warping; gesture learning; gesture recognition; gesture tracking; human gestures; real-time performance; search space pruning; space-time gestures; stereotypical space-time patterns; view prediction; view-based approach; Eyes; Hardware; Humans; Laboratories; Machine vision; Magnetic heads; Pattern matching; Pattern recognition; Predictive models; Statistics;
  • fLanguage
    English
  • Publisher
    ieee
  • Conference_Titel
    Computer Vision and Pattern Recognition, 1993. Proceedings CVPR '93., 1993 IEEE Computer Society Conference on
  • Conference_Location
    New York, NY
  • ISSN
    1063-6919
  • Print_ISBN
    0-8186-3880-X
  • Type

    conf

  • DOI
    10.1109/CVPR.1993.341109
  • Filename
    341109