4.6 Article

Learning Cross-Scale Visual Representations for Real-Time Image Geo-Localization

期刊

IEEE ROBOTICS AND AUTOMATION LETTERS
卷 7, 期 2, 页码 5087-5094

出版社

IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
DOI: 10.1109/LRA.2022.3154035

关键词

Deep learning for visual perception; marine robotics; representation learning

类别

向作者/读者索取更多资源

This study proposes a method to address the challenging task of robot localization in GPS denied environments by localizing image observations in a 2D multimodal geospatial map. The experiments show that the proposed method performs better on smaller-scale multimodal maps, is more computationally efficient for real-time applications, and can be used directly in concert with state estimation pipelines.
Robot localization remains a challenging task in GPS denied environments. State estimation approaches based on local sensors, e.g. cameras or IMUs, are drifting-prone for long-range missions as error accumulates. In this study, we aim to address this problem by localizing image observations in a 2D multimodal geospatial map. We introduce the cross-scale(1) dataset and a methodology to produce additional data from cross-modality sources. We propose a framework that learns cross-scale visual representations without supervision. Experiments are conducted on data from two different domains, underwater and aerial. In contrast to existing studies in cross-view image geo-localization, our approach a) performs better on smaller-scale multi-modal maps; b) is more computationally efficient for real-time applications; c) can serve directly in concert with state estimation pipelines.(2)

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.6
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据