Title :
Joint segmentation and classification of human actions in video
Author :
Hoai, Minh ; Lan, Zhen-Zhong ; De La Torre, Fernando
Author_Institution :
Carnegie Mellon Univ., Pittsburgh, PA, USA
Abstract :
Automatic video segmentation and action recognition has been a long-standing problem in computer vision. Much work in the literature treats video segmentation and action recognition as two independent problems; while segmentation is often done without a temporal model of the activity, action recognition is usually performed on pre-segmented clips. In this paper we propose a novel method that avoids the limitations of the above approaches by jointly performing video segmentation and action recognition. Unlike standard approaches based on extensions of dynamic Bayesian networks, our method is based on a discriminative temporal extension of the spatial bag-of-words model that has been very popular in object recognition. The classification is performed robustly within a multi-class SVM framework whereas the inference over the segments is done efficiently with dynamic programming. Experimental results on honeybee, Weizmann, and Hollywood datasets illustrate the benefits of our approach compared to state-of-the-art methods.
Keywords :
belief networks; computer vision; dynamic programming; image classification; image segmentation; inference mechanisms; object recognition; support vector machines; video signal processing; Hollywood dataset; Weizmann dataset; action recognition; automatic video segmentation; computer vision; discriminative temporal extension; dynamic Bayesian networks; dynamic programming; honeybee dataset; human action classification; human action segmentation; inference; multiclass SVM framework; object recognition; spatial bag-of-words model; Hidden Markov models; Humans; Joints; Motion segmentation; Support vector machines; Time series analysis; Training;
Conference_Titel :
Computer Vision and Pattern Recognition (CVPR), 2011 IEEE Conference on
Conference_Location :
Providence, RI
Print_ISBN :
978-1-4577-0394-2
DOI :
10.1109/CVPR.2011.5995470