首页> 外文期刊>Neurocomputing >ObjectFusion: An object detection and segmentation framework with RGB-D SLAM and convolutional neural networks
【24h】

ObjectFusion: An object detection and segmentation framework with RGB-D SLAM and convolutional neural networks

机译:ObjectFusion:具有RGB-D SLAM和卷积神经网络的对象检测和分割框架

获取原文
获取原文并翻译 | 示例
           

摘要

Given the driving advances on CNNs (Convolutional Neural Networks) [1], deep neural networks being deployed for accurate detection and semantic reconstruction in SLAM (Simultaneous Localization and Mapping) has become a trend. However, as far as we know, almost all existing methods focus on design a specific CNN architecture for single task. In this paper, we propose a novel framework which employs a general object detection CNN to fuse with a SLAM system towards obtaining better performances on both detection and semantic segmentation in 3D space. Our approach first use CNN-based detection network to obtain the 2D object proposals which can be used to establish the local target map. We then use the results estimated from SLAM to update the dynamic global target map based on the local target map obtained by CNNs. Finally, we are able to obtain the detection result for the current frame by projecting the global target map into 2D space. On the other hand, we send the estimation results back to SLAM and update the semantic surfel model in SLAM system. Therefore, we can acquire the segmentation result by projecting the updated 3D surfel model into 2D. Our fusion scheme privileges in object detection and segmentation by integrating with SLAM system to preserve the spatial continuity and temporal consistency. Evaluation performances on four datasets demonstrate the effectiveness and robustness of our method. (C) 2019 Elsevier B.V. All rights reserved.
机译:鉴于CNN(卷积神经网络)[1]的推动进步,在SLAM(同时定位和映射)中部署用于精确检测和语义重构的深度神经网络已成为一种趋势。但是,据我们所知,几乎所有现有方法都专注于为单个任务设计特定的CNN体​​系结构。在本文中,我们提出了一种新颖的框架,该框架采用通用的对象检测CNN与SLAM系统融合,以在3D空间中的检测和语义分割上获得更好的性能。我们的方法首先使用基于CNN的检测网络来获取2D对象建议,这些建议可用于建立本地目标地图。然后,我们使用从SLAM估计的结果,基于CNN获得的本地目标地图来更新动态全局目标地图。最后,通过将全局目标地图投影到2D空间中,我们可以获得当前帧的检测结果。另一方面,我们将估计结果发送回SLAM,并更新SLAM系统中的语义浏览模型。因此,我们可以通过将更新的3D surfel模型投影到2D中来获取分割结果。通过与SLAM系统集成,我们的融合方案优先进行对象检测和分割,以保留空间连续性和时间一致性。在四个数据集上的评估性能证明了我们方法的有效性和鲁棒性。 (C)2019 Elsevier B.V.保留所有权利。

著录项

  • 来源
    《Neurocomputing》 |2019年第14期|3-14|共12页
  • 作者单位

    Zhejiang Univ, Inst Cyber Syst & Control, State Key Lab Ind Control Technol, Hangzhou 310027, Zhejiang, Peoples R China;

    Zhejiang Univ, Inst Cyber Syst & Control, State Key Lab Ind Control Technol, Hangzhou 310027, Zhejiang, Peoples R China;

    Zhejiang Univ, Inst Cyber Syst & Control, State Key Lab Ind Control Technol, Hangzhou 310027, Zhejiang, Peoples R China;

    Kim Il Song Univ, Inst Informat Technol, Pyongyang 190016, South Korea;

    Zhejiang Univ, Inst Cyber Syst & Control, State Key Lab Ind Control Technol, Hangzhou 310027, Zhejiang, Peoples R China;

  • 收录信息 美国《科学引文索引》(SCI);美国《工程索引》(EI);
  • 原文格式 PDF
  • 正文语种 eng
  • 中图分类
  • 关键词

    Intelligent Sensing; CNNs; SLAM; Object detection; Segmentation;

    机译:智能感知;CNN;SLAM;目标检测;分割;

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号