4.7 Article

SVO: Semidirect Visual Odometry for Monocular and Multicamera Systems

Journal

IEEE TRANSACTIONS ON ROBOTICS
Volume 33, Issue 2, Pages 249-265

Publisher

IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
DOI: 10.1109/TRO.2016.2623335

Keywords

Robot vision; simultaneous localization and mapping (SLAM)

Categories

Funding

  1. Swiss National Foundation [200021-143607]
  2. National Center of Competence in Research Robotics
  3. UZH Forschungskredit
  4. SNSF-ERC Starting Grant
  5. Swiss National Science Foundation (SNF) [200021_143607] Funding Source: Swiss National Science Foundation (SNF)

Ask authors/readers for more resources

Direct methods for visual odometry (VO) have gained popularity for their capability to exploit information from all intensity gradients in the image. However, low computational speed as well as missing guarantees for optimality and consistency are limiting factors of direct methods, in which established feature-based methods succeed instead. Based on these considerations, we propose a semidirect VO (SVO) that uses direct methods to track and triangulate pixels that are characterized by high image gradients, but relies on proven feature-based methods for joint optimization of structure and motion. Together with a robust probabilistic depth estimation algorithm, this enables us to efficiently track pixels lying on weak corners and edges in environments with little or high-frequency texture. We further demonstrate that the algorithm can easily be extended to multiple cameras, to track edges, to include motion priors, and to enable the use of very large field of view cameras, such as fisheye and catadioptric ones. Experimental evaluation on benchmark datasets shows that the algorithm is significantly faster than the state of the art while achieving highly competitive accuracy.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.7
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available