Your new experience awaits. Try the new design now and help us make it even better

ORIGINAL RESEARCH article

Front. Robot. AI

Sec. Robot Vision and Artificial Perception

Volume 12 - 2025 | doi: 10.3389/frobt.2025.1644230

This article is part of the Research TopicVision AI in Robotic Perception and MappingView all 3 articles

Dense Mapping From Sparse Visual Odometry: A Lightweight Uncertainty-Guaranteed Depth Completion Method

Provisionally accepted
Daolong  YangDaolong YangXudong  ZhangXudong ZhangHaoyuan  LiuHaoyuan LiuHaoyang  WuHaoyang WuChengcai  WangChengcai Wang*Kun  XuKun Xu*Xilun  DingXilun Ding
  • Beihang University, Beijing, China

The final, formatted version of the article will be published soon.

Visual odometry (VO) has been widely deployed on mobile robots for spatial perception. State-of-the-art VO offers robust localization, the maps it generates are often too sparse for downstream tasks due to insufficient depth data. While depth completion methods can estimate dense depth from sparse data, the extreme sparsity and highly uneven distribution of depth signals in VO (∼ 0.15% of the pixels in the depth image available) poses significant challenges. To address this issue, we propose a lightweight Image-Guided Uncertainty-Aware Depth Completion Network (IU-DC) for completing sparse depth from VO. This network integrates color and spatial information into a normalized convolutional neural network to tackle the sparsity issue and simultaneously outputs dense depth and associated uncertainty. The estimated depth is uncertainty-aware, allowing for the filtering of outliers and ensuring precise spatial perception. The superior performance of IU-DC compared to SOTA is validated across multiple open-source datasets in terms of depth and uncertainty estimation accuracy. In real-world mapping tasks, by integrating IU-DC with the mapping module, we achieve 50× more reconstructed volumes and 78% coverage of the ground truth with twice the accuracy compared to SOTA, despite having only 0.6M parameters (just 3% of the size of the SOTA). Our code will be open-sourced to support future research.

Keywords: Mapping, deep learning for visual perception, visual odometry, Depth Completion ing and Automation, Beihang University, China. {YangDL, sy2207621, liuhaoyuan

Received: 10 Jun 2025; Accepted: 08 Aug 2025.

Copyright: © 2025 Yang, Zhang, Liu, Wu, Wang, Xu and Ding. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) or licensor are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

* Correspondence:
Chengcai Wang, Beihang University, Beijing, China
Kun Xu, Beihang University, Beijing, China

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.