...
首页> 外文期刊>IEEE transactions on multimedia >Real-Time Dense Monocular SLAM With Online Adapted Depth Prediction Network
【24h】

Real-Time Dense Monocular SLAM With Online Adapted Depth Prediction Network

机译:在线自适应深度预测网络的实时密集单眼SLAM

获取原文
获取原文并翻译 | 示例
           

摘要

Considerable advances have been achieved in estimating the depth map from a single image via convolutional neural networks (CNNs) during the past few years. Combining depth prediction from CNNs with conventional monocular simultaneous localization and mapping (SLAM) is promising for accurate and dense monocular reconstruction, in particular addressing the two long-standing challenges in conventional monocular SLAM: low map completeness and scale ambiguity. However, depth estimated by pretrained CNNs usually fails to achieve sufficient accuracy for environments of different types from the training data, which are common for certain applications such as obstacle avoidance of drones in unknown scenes. Additionally, inaccurate depth prediction of CNN could yield large tracking errors in monocular SLAM. In this paper, we present a real-time dense monocular SLAM system, which effectively fuses direct monocular SLAM with an online-adapted depth prediction network for achieving accurate depth prediction of scenes of different types from the training data and providing absolute scale information for tracking and mapping. Specifically, on one hand, tracking pose (i.e., translation and rotation) from direct SLAM is used for selecting a small set of highly effective and reliable training images, which acts as ground truth for tuning the depth prediction network on-the-fly toward better generalization ability for scenes of different types. A stage-wise Stochastic Gradient Descent algorithm with a selective update strategy is introduced for efficient convergence of the tuning process. On the other hand, the dense map produced by the adapted network is applied to address scale ambiguity of direct monocular SLAM which in turn improves the accuracy of both tracking and overall reconstruction. The system with assistance of both CPUs and GPUs, can achieve real-time performance with progressively improved reconstruction accuracy. Experimental results on public datasets and live application to obstacle avoidance of drones demonstrate that our method outperforms the state-of-the-art methods with greater map completeness and accuracy, and a smaller tracking error.
机译:在过去的几年中,在通过卷积神经网络(CNN)从单个图像估计深度图方面已经取得了很大的进步。将CNN的深度预测与常规单眼同时定位和制图(SLAM)相结合,有望实现准确且密集的单眼重建,尤其是解决传统单眼SLAM的两个长期挑战:低地图完整性和比例模糊性。但是,对于来自训练数据的不同类型的环境,由预训练的CNN估计的深度通常无法获得足够的精度,这对于某些应用程序是很常见的,例如在未知场景中避开无人机的障碍。此外,CNN的深度预测不准确会在单眼SLAM中产生较大的跟踪误差。在本文中,我们提出了一种实时密集的单眼SLAM系统,该系统有效地将直接单眼SLAM与在线适应的深度预测网络融合,从而从训练数据中获得不同类型场景的准确深度预测,并提供用于跟踪的绝对比例信息和映射。具体来说,一方面,来自直接SLAM的跟踪姿势(即平移和旋转)用于选择一小组高效且可靠的训练图像,这是用于实时调整深度预测网络的基本事实对不同类型的场景具有更好的泛化能力。引入了具有选择性更新策略的分阶段随机梯度下降算法,可以有效地收敛调整过程。另一方面,由自适应网络产生的密集图被用于解决直接单眼SLAM的比例尺模糊性,这反过来又提高了跟踪和整体重建的准确性。该系统在CPU和GPU的协助下,可以实现实时性能,并逐步提高重建精度。在公共数据集上的实验结果以及在无人驾驶飞机避障方面的实时应用表明,我们的方法以更高的地图完整性和准确性以及较小的跟踪误差优于最新方法。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号