...
首页> 外文期刊>Pattern Recognition: The Journal of the Pattern Recognition Society >Monocular depth estimation with hierarchical fusion of dilated CNNs and soft-weighted-sum inference
【24h】

Monocular depth estimation with hierarchical fusion of dilated CNNs and soft-weighted-sum inference

机译:具有扩张CNN的分层融合的单眼深度估计和软加权 - 和推论

获取原文
获取原文并翻译 | 示例
           

摘要

Monocular depth estimation is very challenging in complex compositions depicting multiple objects of diverse scales. Albeit the recent great progress thanks to the deep convolutional neural networks, the state-of-the-art monocular depth estimation methods still fall short to handle such real-world challenging scenarios. In this paper, we propose a deep end-to-end learning framework to tackle these challenges, which learns the direct mapping from a color image to the corresponding depth map. First, we represent monocular depth estimation as a multi-category dense labeling task by contrast to the regression-based formulation. In this way, we could build upon the recent progress in dense labeling such as semantic segmentation. Second, we fuse different side-outputs from our front-end dilated convolutional neural network in a hierarchical way to exploit the multi-scale depth cues for monocular depth estimation, which is critical in achieving scale-aware depth estimation. Third, we propose to utilize soft-weighted sum inference instead of the hard-max inference, transforming the discretized depth scores to continuous depth values. Thus, we reduce the influence of quantization error and improve the robustness of our method. Extensive experiments have been conducted on the Make3D, NYU v2, and KITTI datasets and superior performance have been achieved on NYU v2 and KITTI datasets compared with current state-of-the-art methods, which shows the superiority of our method. Furthermore, experiments on the NYU v2 dataset reveal that our classification based model is able to learn the probability distribution of depth. (C) 2018 Elsevier Ltd. All rights reserved.
机译:单目深度估计在描绘多个不同尺度物体的复杂构图中非常具有挑战性。尽管最近由于深度卷积神经网络取得了巨大的进步,但最先进的单目深度估计方法仍然无法处理此类现实世界的挑战场景。在本文中,我们提出了一个深度端到端学习框架来应对这些挑战,该框架学习从彩色图像到相应深度图的直接映射。首先,与基于回归的公式相比,我们将单目深度估计表示为多类别密集标记任务。通过这种方式,我们可以在语义分割等密集标记的最新进展的基础上继续发展。其次,我们将前端扩展卷积神经网络的不同侧面输出进行分层融合,以利用多尺度深度线索进行单目深度估计,这对于实现尺度感知深度估计至关重要。第三,我们建议使用软加权和推理代替硬最大推理,将离散化的深度分数转换为连续的深度值。因此,我们减少了量化误差的影响,提高了方法的鲁棒性。在Make3D、NYU v2和KITTI数据集上进行了大量实验,与当前最先进的方法相比,在NYU v2和KITTI数据集上取得了优异的性能,这表明了我们方法的优越性。此外,在纽约大学v2数据集上的实验表明,基于分类的模型能够学习深度的概率分布。(C) 2018爱思唯尔有限公司版权所有。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号