4.5 Article

Encoding multiple contextual clues for partial-duplicate image retrieval

Journal

PATTERN RECOGNITION LETTERS
Volume 109, Issue -, Pages 18-26

Publisher

ELSEVIER
DOI: 10.1016/j.patrec.2017.08.013

Keywords

Image search; Partial-duplicate image retrieval; Near-duplicate image retrieval; Image copy detection; Contextual clue; BOW model

Funding

  1. Canada Research Chair program (CRC)
  2. AUTO21 Network of Centers of Excellence
  3. Natural Sciences and Engineering Research Council of Canada (NSERCC)
  4. National Natural Science Foundation of China [61602253, U1536206, 61232016, U1405254, 61373133, 61502242, 61572258, 61672294]
  5. Jiangsu Basic Research Programs-Natural Science Foundation [BK20150925, BK20151530]
  6. Priority Academic Program Development of Jiangsu Higher Education Institutions (PAPD) fund
  7. Collaborative Innovation Center of Atmospheric Environment and Equipment Technology (CICAEET) fund

Ask authors/readers for more resources

Recently, most of partial-duplicate image retrieval approaches build on the bag-of-visual-words (BOW) model, in which local image features are quantized into a bag of compact visual words, i.e., BOW representation, for fast image matching. However, due to the quantization errors in visual words, the BOW representation shows low discriminability, which causes negative influence on retrieval accuracy. Encoding contextual clues into the BOW representation is a popular technique to improve its discriminability. Unfortunately, the captured contextual clues are generally not stable and informative enough, resulting in limited discriminability improvement. To address the issues, we propose a multiple contextual clue encoding approach for partial-duplicate image retrieval. By treating each visual word of any given query or database image as a center, we first propose an asymmetrical context selection strategy to select the contextual visual words for the query and database images differently. Then, we capture the multiple contextual clues: the geometric relationships, the visual relationships, and the spatial configurations between the center and its contextual visual words. These captured multiple contextual clues are compressed to generate the multi-contextual descriptors, which are further integrated with the center visual word to improve the discriminability of BOW representation. Experiments conducted on the large-scale partial-duplicate image dataset demonstrate that the proposed approach provides higher retrieval accuracy than the state-of-the-arts, while achieves comparable performances in time and space efficiency. (C) 2017 Published by Elsevier B.V.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.5
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available