ORIGINAL RESEARCH article
Front. Robot. AI
Sec. Robot Vision and Artificial Perception
Volume 12 - 2025 | doi: 10.3389/frobt.2025.1644230
This article is part of the Research TopicVision AI in Robotic Perception and MappingView all 3 articles
Dense Mapping From Sparse Visual Odometry: A Lightweight Uncertainty-Guaranteed Depth Completion Method
Provisionally accepted- Beihang University, Beijing, China
Select one of your emails
You have multiple emails registered with Frontiers:
Notify me on publication
Please enter your email address:
If you already have an account, please login
You don't have a Frontiers account ? You can register here
Visual odometry (VO) has been widely deployed on mobile robots for spatial perception. State-of-the-art VO offers robust localization, the maps it generates are often too sparse for downstream tasks due to insufficient depth data. While depth completion methods can estimate dense depth from sparse data, the extreme sparsity and highly uneven distribution of depth signals in VO (∼ 0.15% of the pixels in the depth image available) poses significant challenges. To address this issue, we propose a lightweight Image-Guided Uncertainty-Aware Depth Completion Network (IU-DC) for completing sparse depth from VO. This network integrates color and spatial information into a normalized convolutional neural network to tackle the sparsity issue and simultaneously outputs dense depth and associated uncertainty. The estimated depth is uncertainty-aware, allowing for the filtering of outliers and ensuring precise spatial perception. The superior performance of IU-DC compared to SOTA is validated across multiple open-source datasets in terms of depth and uncertainty estimation accuracy. In real-world mapping tasks, by integrating IU-DC with the mapping module, we achieve 50× more reconstructed volumes and 78% coverage of the ground truth with twice the accuracy compared to SOTA, despite having only 0.6M parameters (just 3% of the size of the SOTA). Our code will be open-sourced to support future research.
Keywords: Mapping, deep learning for visual perception, visual odometry, Depth Completion ing and Automation, Beihang University, China. {YangDL, sy2207621, liuhaoyuan
Received: 10 Jun 2025; Accepted: 08 Aug 2025.
Copyright: © 2025 Yang, Zhang, Liu, Wu, Wang, Xu and Ding. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) or licensor are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
* Correspondence:
Chengcai Wang, Beihang University, Beijing, China
Kun Xu, Beihang University, Beijing, China
Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.