4.7 Article

Improving Image Description with Auxiliary Modality for Visual Localization in Challenging Conditions

Journal

INTERNATIONAL JOURNAL OF COMPUTER VISION
Volume 129, Issue 1, Pages -

Publisher

SPRINGER
DOI: 10.1007/s11263-020-01363-6

Keywords

Localization; Image retrieval; Side modality learning; Depth from monocular; Global image descriptor

Funding

  1. French ANR project pLaTINUM [ANR-15-CE23-0010]
  2. NVIDIA Corporation
  3. Agence Nationale de la Recherche (ANR) [ANR-15-CE23-0010] Funding Source: Agence Nationale de la Recherche (ANR)

Ask authors/readers for more resources

The study introduces a new outdoor large-scale image-based localization method with powerful cross-season, cross-weather, and day/night localization capabilities. By learning global image descriptors and combining scene geometry information, the method enhances localization accuracy. It effectively utilizes visual and geometric cues from monocular images, performing well in challenging localization scenarios.
Image indexing for lifelong localization is a key component for a large panel of applications, including robot navigation, autonomous driving or cultural heritage valorization. The principal difficulty in long-term localization arises from the dynamic changes that affect outdoor environments. In this work, we propose a new approach for outdoor large scale image-based localization that can deal with challenging scenarios like cross-season, cross-weather and day/night localization. The key component of our method is a new learned global image descriptor, that can effectively benefit from scene geometry information during training. At test time, our system is capable of inferring the depth map related to the query image and use it to increase localization accuracy. We show through extensive evaluation that our method can improve localization performances, especially in challenging scenarios when the visual appearance of the scene has changed. Our method is able to leverage both visual and geometric clues from monocular images to create discriminative descriptors for cross-season localization and effective matching of images acquired at different time periods. Our method can also use weakly annotated data to localize night images across a reference dataset of daytime images. Finally we extended our method to reflectance modality and we compare multi-modal descriptors respectively based on geometry, material reflectance and a combination of both.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.7
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available