Title :
Top-down visual saliency via joint CRF and dictionary learning
Author :
Yang, Jimei ; Yang, Ming-Hsuan
Author_Institution :
Univ. of California at Merced, Merced, CA, USA
Abstract :
Top-down visual saliency facilities object localization by providing a discriminative representation of target objects and a probability map for reducing the search space. In this paper, we propose a novel top-down saliency model that jointly learns a Conditional Random Field (CRF) and a discriminative dictionary. The proposed model is formulated based on a CRF with latent variables. By using sparse codes as latent variables, we train the dictionary modulated by CRF, and meanwhile a CRF with sparse coding. We propose a max-margin approach to train our model via fast inference algorithms. We evaluate our model on the Graz-02 and PASCAL VOC 2007 datasets. Experimental results show that our model performs favorably against the state-of-the-art top-down saliency methods. We also observe that the dictionary update significantly improves the model performance.
Keywords :
inference mechanisms; object detection; object recognition; probability; random processes; conditional random field; dictionary learning; discriminative dictionary; discriminative representation; inference algorithm; joint CRF; latent variables; max-margin approach; probability map; search space; sparse codes; sparse coding; target objects; top-down saliency method; top-down saliency model; top-down visual saliency facilities object localization; Bicycles; Computational modeling; Dictionaries; Encoding; Joints; Training; Visualization;
Conference_Titel :
Computer Vision and Pattern Recognition (CVPR), 2012 IEEE Conference on
Conference_Location :
Providence, RI
Print_ISBN :
978-1-4673-1226-4
Electronic_ISBN :
1063-6919
DOI :
10.1109/CVPR.2012.6247940