• DocumentCode
    2266640
  • Title

    Trajectons: Action recognition through the motion analysis of tracked features

  • Author

    Matikainen, Pyry ; Hebert, Martial ; Sukthankar, Rahul

  • Author_Institution
    Robot. Inst., Carnegie Mellon Univ., Pittsburgh, PA, USA
  • fYear
    2009
  • fDate
    Sept. 27 2009-Oct. 4 2009
  • Firstpage
    514
  • Lastpage
    521
  • Abstract
    The defining feature of video compared to still images is motion, and as such the selection of good motion features for action recognition is crucial, especially for bag of words techniques that rely heavily on their features. Existing motion techniques either assume that a difficult problem like background/foreground segmentation has already been solved (contour/silhouette based techniques) or are computationally expensive and prone to noise (optical flow). We present a technique for motion based on quantized trajectory snippets of tracked features. These quantized snippets, or trajectons, rely only on simple feature tracking and are computationally efficient. We demonstrate that within a bag of words framework trajectons can match state of the art results, slightly outperforming histogram of optical flow features on the Hollywood Actions dataset. Additionally, we present qualitative results in a video search task on a custom dataset of challenging YouTube videos.
  • Keywords
    image segmentation; image sequences; motion estimation; video retrieval; YouTube videos; action recognition; background-foreground segmentation; hollywood actions dataset; optical flow; tracked features motion analysis; trajectons; video defining feature; video search task; Background noise; Histograms; Image motion analysis; Image recognition; Image segmentation; Motion analysis; Optical computing; Optical noise; Tracking; Trajectory;
  • fLanguage
    English
  • Publisher
    ieee
  • Conference_Titel
    Computer Vision Workshops (ICCV Workshops), 2009 IEEE 12th International Conference on
  • Conference_Location
    Kyoto
  • Print_ISBN
    978-1-4244-4442-7
  • Electronic_ISBN
    978-1-4244-4441-0
  • Type

    conf

  • DOI
    10.1109/ICCVW.2009.5457659
  • Filename
    5457659