首页> 外文会议>International Symposium on Experimental Robotics >Deep Multispectral Semantic Scene Understanding of Forested Environments Using Multimodal Fusion
【24h】

Deep Multispectral Semantic Scene Understanding of Forested Environments Using Multimodal Fusion

机译:使用多模式融合的深度多光谱语义场景理解森林环境

获取原文

摘要

Semantic scene understanding of unstructured environments is a highly challenging task for robots operating in the real world. Deep Convolutional Neural Network architectures define the state of the art in various segmentation tasks. So far, researchers have focused on segmentation with RGB data. In this paper, we study the use of multispectral and multimodal images for semantic segmentation and develop fusion architectures that learn from RGB, Near-InfraRed channels, and depth data. We introduce a first-of-its-kind multispectral segmentation benchmark that contains 15,000 images and 366 pixel-wise ground truth annotations of unstructured forest environments. We identify new data augmentation strategies that enable training of very deep models using relatively small datasets. We show that our UpNet architecture exceeds the state of the art both qualitatively and quantitatively on our benchmark. In addition, we present experimental results for segmentation under challenging real-world conditions. Benchmark and demo are publicly available at http://deepscene.cs.uni-freiburg.de.
机译:对非结构化环境的语义现场了解是一个高度挑战的现实世界的机器人任务。深度卷积神经网络架构在各种分割任务中定义了本领域的状态。到目前为止,研究人员专注于RGB数据的分割。在本文中,我们研究了使用多级和多模式图像进行语义分割和开发学习的融合架构,从RGB,近红外通道和深度数据中学习。我们介绍了一系列的多光谱分割基准,其中包含15,000个图像和366个像素 - 明智的非结构化林环境的实践注释。我们确定使用相对较小的数据集启用非常深模型的新数据增强策略。我们表明,我们的UPNet架构在我们的基准测试中定制和定量地超出了最先进的艺术状态。此外,我们提出了在挑战现实世界条件下的分割实验结果。基准和演示在http://deepscene.cs.uni-freiburg.de上公开提供。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号