4.5 Article

Encoding multiple contextual clues for partial-duplicate image retrieval

期刊

PATTERN RECOGNITION LETTERS
卷 109, 期 -, 页码 18-26

出版社

ELSEVIER
DOI: 10.1016/j.patrec.2017.08.013

关键词

Image search; Partial-duplicate image retrieval; Near-duplicate image retrieval; Image copy detection; Contextual clue; BOW model

资金

  1. Canada Research Chair program (CRC)
  2. AUTO21 Network of Centers of Excellence
  3. Natural Sciences and Engineering Research Council of Canada (NSERCC)
  4. National Natural Science Foundation of China [61602253, U1536206, 61232016, U1405254, 61373133, 61502242, 61572258, 61672294]
  5. Jiangsu Basic Research Programs-Natural Science Foundation [BK20150925, BK20151530]
  6. Priority Academic Program Development of Jiangsu Higher Education Institutions (PAPD) fund
  7. Collaborative Innovation Center of Atmospheric Environment and Equipment Technology (CICAEET) fund

向作者/读者索取更多资源

Recently, most of partial-duplicate image retrieval approaches build on the bag-of-visual-words (BOW) model, in which local image features are quantized into a bag of compact visual words, i.e., BOW representation, for fast image matching. However, due to the quantization errors in visual words, the BOW representation shows low discriminability, which causes negative influence on retrieval accuracy. Encoding contextual clues into the BOW representation is a popular technique to improve its discriminability. Unfortunately, the captured contextual clues are generally not stable and informative enough, resulting in limited discriminability improvement. To address the issues, we propose a multiple contextual clue encoding approach for partial-duplicate image retrieval. By treating each visual word of any given query or database image as a center, we first propose an asymmetrical context selection strategy to select the contextual visual words for the query and database images differently. Then, we capture the multiple contextual clues: the geometric relationships, the visual relationships, and the spatial configurations between the center and its contextual visual words. These captured multiple contextual clues are compressed to generate the multi-contextual descriptors, which are further integrated with the center visual word to improve the discriminability of BOW representation. Experiments conducted on the large-scale partial-duplicate image dataset demonstrate that the proposed approach provides higher retrieval accuracy than the state-of-the-arts, while achieves comparable performances in time and space efficiency. (C) 2017 Published by Elsevier B.V.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.5
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据