首页> 外文会议>IEEE International Conference on Robotics and Automation >Fusion of Stereo and Still Monocular Depth Estimates in a Self-Supervised Learning Context
【24h】

Fusion of Stereo and Still Monocular Depth Estimates in a Self-Supervised Learning Context

机译:立体声融合和静态深度估计在自我监督的学习背景下

获取原文

摘要

We study how autonomous robots can learn by themselves to improve their depth estimation capability. In particular, we investigate a self-supervised learning setup in which stereo vision depth estimates serve as targets for a convolutional neural network (CNN) that transforms a single still image to a dense depth map. After training, the stereo and mono estimates are fused with a novel fusion method that preserves high confidence stereo estimates, while leveraging the CNN estimates in the low-confidence regions. The main contribution of the article is that it is shown that the fused estimates lead to a higher performance than the stereo vision estimates alone. Experiments are performed on the KITTI dataset, and on board of a Parrot SLAMDunk, showing that even rather limited CNNs can help provide stereo vision equipped robots with more reliable depth maps for autonomous navigation.
机译:我们研究自主机器人可以自行学习如何提高他们的深度估计能力。特别是,我们调查一个自我监督的学习设置,其中立体视觉深度估计用作卷积神经网络(CNN)的目标,其将单个静止图像转换为密集的深度图。在培训之后,立体声和单声道估计与一种新的融合方法融合,这些方法保留了高置信立体声估计,同时利用低置信区中的CNN估计。物品的主要贡献是,表明融合估计率远远超过立体视觉估计的性能。实验在Kitti DataSet上进行,并在鹦鹉陷阱船上进行,表明即使相当有限的CNN可以帮助提供配备的立体视觉的机器人,为自主导航提供更可靠的深度图。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号