• DocumentCode
    72927
  • Title

    Visual Phraselet: Refining Spatial Constraints for Large Scale Image Search

  • Author

    Liang Zheng ; Shengjin Wang

  • Author_Institution
    Dept. of Electron. Eng., Tsinghua Univ., Beijing, China
  • Volume
    20
  • Issue
    4
  • fYear
    2013
  • fDate
    Apr-13
  • Firstpage
    391
  • Lastpage
    394
  • Abstract
    The Bag-of-Words (BoW) model is prone to the deficiency of spatial constraints among visual words. The state of the art methods encode spatial information via visual phrases. However, these methods discard the spatial context among visual phrases instead. To address the problem, this letter introduces a novel visual concept, the Visual Phraselet, as a kind of similarity measurement between images. The visual phraselet refers to the spatial consistent group of visual phrases. In a simple yet effective manner, visual phraselet filters out false visual phrase matches, and is much more discriminative than both visual word and visual phrase. To boost the discovery of visual phraselets, we apply the soft quantization scheme. Our method is evaluated through extensive experiments on three benchmark datasets (Oxford 5 K, Paris 6 K and Flickr 1 M). We report significant improvements as large as 54.6% over the baseline approach, thus validating the concept of visual phraselet.
  • Keywords
    image coding; image matching; image retrieval; quantisation (signal); Visual Phraselet; bag-of-words model; large scale image search; refining spatial constraints; similarity measurement; soft quantization scheme; spatial context; spatial information encoding; visual phraselet filters; visual phraselets; visual phrases; visual words; Image search; spatial constraint; visual phrase; visual phraselet;
  • fLanguage
    English
  • Journal_Title
    Signal Processing Letters, IEEE
  • Publisher
    ieee
  • ISSN
    1070-9908
  • Type

    jour

  • DOI
    10.1109/LSP.2013.2249513
  • Filename
    6471751