Abstract

We present an approach to depth estimation that fuses information from a stereo pair with sparse range measurements derived from a LIDAR sensor or a range camera. The goal of this work is to exploit the complementary strengths of the two sensor modalities, the accurate but sparse range measurements and the ambiguous but dense stereo information. These two sources are effectively and efficiently fused by combining ideas from anisotropic diffusion and semi-global matching. We evaluate our approach on the KITTI 2015 and Middlebury 2014 datasets, using randomly sampled ground truth range measurements as our sparse depth input. We achieve significant performance improvements with a small fraction of range measurements on both datasets. We also provide qualitative results from our platform using the PMDTec Monstar sensor. Our entire pipeline runs on an NVIDIA TX-2 platform at 5Hz on 1280x1024 stereo images with 128 disparity levels.

Comment: 7 pages, 5 figures, 2 table


Original document

The different versions of the original document can be found in:

http://dx.doi.org/10.1109/icra.2019.8794023
https://dblp.uni-trier.de/db/conf/icra/icra2019.html#ShivakumarMPKT19,
https://ieeexplore.ieee.org/abstract/document/8794023,
https://arxiv.org/abs/1809.07677,
https://www.arxiv-vanity.com/papers/1809.07677,
https://academic.microsoft.com/#/detail/2967922036
Back to Top

Document information

Published on 01/01/2018

Volume 2018, 2018
DOI: 10.1109/icra.2019.8794023
Licence: CC BY-NC-SA license

Document Score

0

Views 5
Recommendations 0

Share this document

claim authorship

Are you one of the authors of this document?