4.7 Article

Salient-VPR: Salient Weighted Global Descriptor for Visual Place Recognition

出版社

IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
DOI: 10.1109/TIM.2022.3219466

关键词

Computer vision; feature representation; image retrieval; place recognition; saliency prediction; semantic segmentation

资金

  1. Collaborative Innovation Center for Deep Space Exploration [SKTC202204]
  2. National Natural Science Foundation of China [51605054]

向作者/读者索取更多资源

This article proposes a salient visual place recognition method that combines image retrieval, semantic information, and saliency cues for accurate estimations. The method uses a novel formulation to combine local semantic features into global descriptors and predicts saliency descriptors to learn the representation of static objects. A late-fusion module is introduced to increase the stability of the descriptor.
Visual place recognition (VPR) is a widely investigated but challenging problem, which must deal with the twin problems of changed appearance caused by varying weather conditions, illumination, and seasons, as well as dynamic objects in complex environments. In this article, we proposed salient VPR by combining image retrieval, semantic information, and saliency cues to achieve accurate estimations. We provide a novel formulation for combining the local semantic features into global descriptors. Unlike traditional global descriptors that typically summarize the whole visual content in the images or image regions, we aggregate dense local semantic descriptors based on pixel-level semantic scores to form global semantic descriptors. Then we calculate predicted saliency descriptors to learn the representation of static objects from images. These salient descriptions are learned from a dataset tailored for VPR missions. Finally, we introduce a late-fusion module, which increases the stability of the descriptor and avoids performance degradation caused by the limitations of semantic and saliency prediction. Our method outperforms traditional global descriptors in experiments, attaining state-of-the-art visual location recognition scores on a variety of challenging datasets such as Pitts30 k, Mapillary Street-level Sequences (MSLS), and Tokyo24/7.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.7
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据