...
【24h】

DEEP LEARNING FOR MONOCULAR DEPTH ESTIMATION FROM UAV IMAGES

机译:从UAV图像中的单眼深度估计深度学习

获取原文
           

摘要

Depth is an essential component for various scene understanding tasks and for reconstructing the 3D geometry of the scene. Estimating depth from stereo images requires multiple views of the same scene to be captured which is often not possible when exploring new environments with a UAV. To overcome this monocular depth estimation has been a topic of interest with the recent advancements in computer vision and deep learning techniques. This research has been widely focused on indoor scenes or outdoor scenes captured at ground level. Single image depth estimation from aerial images has been limited due to additional complexities arising from increased camera distance, wider area coverage with lots of occlusions. A new aerial image dataset is prepared specifically for this purpose combining Unmanned Aerial Vehicles (UAV) images covering different regions, features and point of views. The single image depth estimation is based on image reconstruction techniques which uses stereo images for learning to estimate depth from single images. Among the various available models for ground-level single image depth estimation, two models, 1) a Convolutional Neural Network (CNN) and 2) a Generative Adversarial model (GAN) are used to learn depth from aerial images from UAVs. These models generate pixel-wise disparity images which could be converted into depth information. The generated disparity maps from these models are evaluated for its internal quality using various error metrics. The results show higher disparity ranges with smoother images generated by CNN model and sharper images with lesser disparity range generated by GAN model. The produced disparity images are converted to depth information and compared with point clouds obtained using Pix4D. It is found that the CNN model performs better than GAN and produces depth similar to that of Pix4D. This comparison helps in streamlining the efforts to produce depth from a single aerial image.
机译:深度是各种场景理解任务的重要组成部分,以及重建场景的3D几何形状。估计来自立体图像的深度需要要捕获的相同场景的多个视图,这通常是不可能的,当探索具有UAV的新环境时。为了克服这种单眼深度估计是计算机视觉和深度学习技术的最新进步感兴趣的主题。该研究已广泛关注地面捕获的室内场景或室外场景。由于摄像机距离增加,因此由于相机距离增加,并且具有许多闭塞而导致的额外复杂性,从航拍图像的单个图像深度估计受到限制。为此目的,特别是为此目的准备了一种新的空中图像数据集,这些目的结合了覆盖不同地区的无人机(UAV)图像,特征和视点。单个图像深度估计基于图像重建技术,该技术使用立体图像来学习从单个图像估计深度。在地面单个图像深度估计的各种可用模型中,两个模型,1)卷积神经网络(CNN)和2)用于从UAVS的空中图像学习深度。这些模型生成可以转换为深度信息的像素 - WISE视差图像。使用各种错误指标评估来自这些模型的产生的差异图。结果显示了由CNN模型产生的更高的视差范围,并通过GaN模型产生的较低的差距范围更高的图像。产生的视差图像被转换为​​深度信息,并与使用PIX4D获得的点云进行比较。发现CNN模型比GaN更好地执行,并产生与PIX4D类似的深度。这种比较有助于简化从单个航拍图像产生深度的努力。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号