首页> 外文会议>International Joint Conference on Computer Vision, Imaging and Computer Graphics Theory and Applications >Exploration of Deep Learning-based Multimodal Fusion for Semantic Road Scene Segmentation
【24h】

Exploration of Deep Learning-based Multimodal Fusion for Semantic Road Scene Segmentation

机译:基于深度学习的多峰融合探讨对语义路面分割的影响

获取原文

摘要

Deep neural networks have been frequently used for semantic scene understanding in recent years. Effective and robust segmentation in outdoor scene is prerequisite for safe autonomous navigation of autonomous vehicles. In this paper, our aim is to find the best exploitation of different imaging modalities for road scene segmentation, as opposed to using a single RGB modality. We explore deep learning-based early and later fusion pattern for semantic segmentation, and propose a new multi-level feature fusion network. Given a pair of aligned multimodal images, the network can achieve faster convergence and incorporate more contextual information. In particular, we introduce the first-of-its-kind dataset, which contains aligned raw RGB images and polarimetric images, followed by manually labeled ground truth. The use of polarization cameras is a sensory augmentation that can significantly enhance the capabilities of image understanding, for the detection of highly reflective areas such as glasses and water. Experimental results suggest that our proposed multimodal fusion network outperforms unimodal networks and two typical fusion architectures.
机译:近年来,深神经网络经常用于语义场景理解。户外场景中的有效和强大的细分是自主车辆安全自主导航的先决条件。在本文中,我们的目标是找到对道路场景分割的不同成像模式的最佳开发,而不是使用单个RGB模态。我们探索基于深度学习的早期和以后的融合模式,用于语义分割,并提出了一种新的多级特征融合网络。给定一对对齐的多模式图像,网络可以实现更快的收敛并结合更多的上下文信息。特别是,我们介绍了一系列的数据集,其中包含对齐的原始RGB图像和偏振图像,然后是手动标记的地面真理。偏振相机的使用是一种感觉增强,可以显着提高图像理解的能力,用于检测高度反射区域,例如眼镜和水。实验结果表明,我们提出的多模融合网络优于单向网络和两个典型的融合架构。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号