首页> 外文会议>IEEE Applied Imagery Pattern Recognition Workshop >Multi-modal Data Analysis and Fusion for Robust Object Detection in 2D/3D Sensing
【24h】

Multi-modal Data Analysis and Fusion for Robust Object Detection in 2D/3D Sensing

机译:2D / 3D感测中鲁棒对象检测的多模态数据分析和融合

获取原文

摘要

Multi-modal data is useful for complex imaging scenarios due to the exclusivity of information found in each modality, but there is a lack of meaningful comparisons of different modalities for object detection. In our work, we propose three contributions: (1) Release of a multi-modal, ground-based small object detection dataset, (2) A performance comparison of 2D and 3D imaging modalities using state-of-the-art algorithms, and (3) a multi-modal fusion framework for 2D/3D sensing. The new dataset encompasses various small objects for detection in EO, IR, and LiDAR modalities. The labeled data has comparable resolutions across each modality for better performance analysis. The modality comparison conducted in this work uses advanced deep learning algorithms, such as Mask R-CNN for 2D imaging and PointNet++ for 3D imaging. The comparisons are conducted with similar parameter sizes and the results are analyzed for specific instances where each modality performed the best. To complement the effectiveness of different data modalities, we developed a fusion strategy to combine detection networks operating on different modalities into a single detection output for accurate object detection and region segmentation. Our fusion strategy utilized the state of the art networks listed above as backbone networks to obtain a confidence score from each modality. The network then determines which modality to base the object detection off of based on those confidences. The effectiveness of the proposed fusion method is being evaluated on the multi-modal dataset for object detection and segmentation and we observe superior performance when compared to single-modality algorithms.
机译:由于每个码形中发现的信息的排他性,多模态数据对于复杂的成像方案非常有用,但是对物体检测的不同模式缺乏有意义的比较。在我们的工作中,我们提出了三个贡献:(1)释放多模态,基于地面小物体检测数据集,(2)使用最先进的算法的2D和3D成像方式的性能比较,以及(3)用于2D / 3D感测的多模态融合框架。新数据集包含各种小型物体,用于在EO,IR和LIDAR方式中检测。标记数据在每个模态中具有可比的分辨率,以进行更好的性能分析。在本工作中进行的模态比较使用高级深度学习算法,例如用于2D成像和PointNet ++的掩模R-CNN,用于3D成像。使用相似的参数大小进行比较,并对每个模态执行最佳的特定实例进行分析结果。为了补充不同数据模式的有效性,我们开发了一种将在不同方式运行的检测网络与精确对象检测和区域分割的单个检测输出相结合的融合策略。我们的融合策略利用上面列出的美术网络的状态作为骨干网络,以获得来自每种方式的置信度。然后,网络确定基于这些信心的对象检测到哪个模块。在与单模算法相比,正在对对象检测和分割的多模态数据集进行评估所提出的融合方法的有效性,并且在单模态算法相比,我们观察卓越的性能。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号