首页> 外文会议>Unconventional and Indirect Imaging, Image Reconstruction, and Wavefront Sensing 2017 >High resolution depth reconstruction from monocular images and sparse point clouds using deep convolutional neural network
【24h】

High resolution depth reconstruction from monocular images and sparse point clouds using deep convolutional neural network

机译:使用深度卷积神经网络从单眼图像和稀疏点云进行高分辨率深度重构

获取原文
获取原文并翻译 | 示例

摘要

Understanding the 3D structure of the environment is advantageous for many tasks in the field of robotics and autonomous vehicles. From the robot's point of view, 3D perception is often formulated as a depth image reconstruction problem. In the literature, dense depth images are often recovered deterministically from stereo image disparities. Other systems use an expensive LiDAR sensor to produce accurate, but semi-sparse depth images. With the advent of deep learning there have also been attempts to estimate depth by only using monocular images. In this paper we combine the best of the two worlds, focusing on a combination of monocular images and low cost LiDAR point clouds. We explore the idea that very sparse depth information accurately captures the global scene structure while variations in image patches can be used to reconstruct local depth to a high resolution. The main contribution of this paper is a supervised learning depth reconstruction system based on a deep convolutional neural network. The network is trained on RGB image patches reinforced with sparse depth information and the output is a depth estimate for each pixel. Using image and point cloud data from the KITTI vision dataset we are able to learn a correspondence between local RGB information and local depth, while at the same time preserving the global scene structure. Our results are evaluated on sequences from the KITTI dataset and our own recordings using a low cost camera and LiDAR setup.
机译:了解环境的3D结构对于机器人技术和自动驾驶汽车领域的许多任务非常有利。从机器人的角度来看,3D感知通常被表述为深度图像重建问题。在文献中,通常从立体图像视差确定性地恢复密集深度图像。其他系统使用昂贵的LiDAR传感器来生成准确但半稀疏的深度图像。随着深度学习的到来,也已经尝试仅通过使用单眼图像来估计深度。在本文中,我们结合了两个方面的优势,重点放在单眼图像和低成本LiDAR点云的结合上。我们探索这样的想法,即非常稀疏的深度信息可以准确地捕获全局场景结构,而图像补丁中的变化可用于将局部深度重构为高分辨率。本文的主要贡献是基于深度卷积神经网络的监督学习深度重建系统。在经过稀疏深度信息增强的RGB图像斑块上训练网络,并且输出是每个像素的深度估计。使用KITTI视觉数据集中的图像和点云数据,我们可以了解局部RGB信息和局部深度之间的对应关系,同时保留全局场景结构。我们使用KITTI数据集的序列对结果进行评估,并使用低成本相机和LiDAR设置对我们自己的记录进行了评估。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号