DocumentCode :
639575
Title :
MODEC: Multimodal Decomposable Models for Human Pose Estimation
Author :
Sapp, Brian ; Taskar, Ben
fYear :
2013
fDate :
23-28 June 2013
Firstpage :
3674
Lastpage :
3681
Abstract :
We propose a multimodal, decomposable model for articulated human pose estimation in monocular images. A typical approach to this problem is to use a linear structured model, which struggles to capture the wide range of appearance present in realistic, unconstrained images. In this paper, we instead propose a model of human pose that explicitly captures a variety of pose modes. Unlike other multimodal models, our approach includes both global and local pose cues and uses a convex objective and joint training for mode selection and pose estimation. We also employ a cascaded mode selection step which controls the trade-off between speed and accuracy, yielding a 5x speedup in inference and learning. Our model outperforms state-of-the-art approaches across the accuracy-speed trade-off curve for several pose datasets. This includes our newly-collected dataset of people in movies, FLIC, which contains an order of magnitude more labeled data for training and testing than existing datasets.
Keywords :
convex programming; image motion analysis; pose estimation; MODEC; articulated human pose estimation; convex objective; global pose cues; linear structured model; local pose cues; monocular images; multimodal decomposable models for human pose estimation; unconstrained images; Computational modeling; Equations; Estimation; Joints; Mathematical model; Training; Zirconium;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Computer Vision and Pattern Recognition (CVPR), 2013 IEEE Conference on
Conference_Location :
Portland, OR
ISSN :
1063-6919
Type :
conf
DOI :
10.1109/CVPR.2013.471
Filename :
6619315
Link To Document :
بازگشت