Real-time Depth Enhanced Monocular Odometry - Robotics Institute Carnegie Mellon University

Real-time Depth Enhanced Monocular Odometry

Conference Paper, Proceedings of (IROS) IEEE/RSJ International Conference on Intelligent Robots and Systems, pp. 4973 - 4980, September, 2014

Abstract

Visual odometry can be augmented by depth information such as provided by RGB-D cameras, or from lidars associated with cameras. However, such depth information can be limited by the sensors, leaving large areas in the visual images where depth is unavailable. Here, we propose a method to utilize the depth, even if sparsely available, in recovery of camera motion. In addition, the method utilizes depth by triangulation from the previously estimated motion, and salient visual features for which depth is unavailable. The core of our method is a bundle adjustment that refines the motion estimates in parallel by processing a sequence of images, in a batch optimization. We have evaluated our method in three sensor setups, one using an RGB-D camera, and two using combinations of a camera and a 3D lidar. Our method is rated #2 on the KITTI odometry benchmark irrespective of sensing modality, and is rated #1 among visual odometry methods.

BibTeX

@conference{Zhang-2014-7928,
author = {Ji Zhang and Michael Kaess and Sanjiv Singh},
title = {Real-time Depth Enhanced Monocular Odometry},
booktitle = {Proceedings of (IROS) IEEE/RSJ International Conference on Intelligent Robots and Systems},
year = {2014},
month = {September},
pages = {4973 - 4980},
}