DocumentCode
72927
Title
Visual Phraselet: Refining Spatial Constraints for Large Scale Image Search
Author
Liang Zheng ; Shengjin Wang
Author_Institution
Dept. of Electron. Eng., Tsinghua Univ., Beijing, China
Volume
20
Issue
4
fYear
2013
fDate
Apr-13
Firstpage
391
Lastpage
394
Abstract
The Bag-of-Words (BoW) model is prone to the deficiency of spatial constraints among visual words. The state of the art methods encode spatial information via visual phrases. However, these methods discard the spatial context among visual phrases instead. To address the problem, this letter introduces a novel visual concept, the Visual Phraselet, as a kind of similarity measurement between images. The visual phraselet refers to the spatial consistent group of visual phrases. In a simple yet effective manner, visual phraselet filters out false visual phrase matches, and is much more discriminative than both visual word and visual phrase. To boost the discovery of visual phraselets, we apply the soft quantization scheme. Our method is evaluated through extensive experiments on three benchmark datasets (Oxford 5 K, Paris 6 K and Flickr 1 M). We report significant improvements as large as 54.6% over the baseline approach, thus validating the concept of visual phraselet.
Keywords
image coding; image matching; image retrieval; quantisation (signal); Visual Phraselet; bag-of-words model; large scale image search; refining spatial constraints; similarity measurement; soft quantization scheme; spatial context; spatial information encoding; visual phraselet filters; visual phraselets; visual phrases; visual words; Image search; spatial constraint; visual phrase; visual phraselet;
fLanguage
English
Journal_Title
Signal Processing Letters, IEEE
Publisher
ieee
ISSN
1070-9908
Type
jour
DOI
10.1109/LSP.2013.2249513
Filename
6471751
Link To Document