首页> 美国卫生研究院文献>Sensors (Basel Switzerland) >Transfer Learning Based Semantic Segmentation for 3D Object Detection from Point Cloud
【2h】

Transfer Learning Based Semantic Segmentation for 3D Object Detection from Point Cloud

机译:从点云传输3D对象检测的基于学习的语义分割

代理获取
本网站仅为用户提供外文OA文献查询和代理获取服务,本网站没有原文。下单后我们将采用程序或人工为您竭诚获取高质量的原文,但由于OA文献来源多样且变更频繁,仍可能出现获取不到、文献不完整或与标题不符等情况,如果获取不到我们将提供退款服务。请知悉。

摘要

Three-dimensional object detection utilizing LiDAR point cloud data is an indispensable part of autonomous driving perception systems. Point cloud-based 3D object detection has been a better replacement for higher accuracy than cameras during nighttime. However, most LiDAR-based 3D object methods work in a supervised manner, which means their state-of-the-art performance relies heavily on a large-scale and well-labeled dataset, while these annotated datasets could be expensive to obtain and only accessible in the limited scenario. Transfer learning is a promising approach to reduce the large-scale training datasets requirement, but existing transfer learning object detectors are primarily for 2D object detection rather than 3D. In this work, we utilize the 3D point cloud data more effectively by representing the birds-eye-view (BEV) scene and propose a transfer learning based point cloud semantic segmentation for 3D object detection. The proposed model minimizes the need for large-scale training datasets and consequently reduces the training time. First, a preprocessing stage filters the raw point cloud data to a BEV map within a specific field of view. Second, the transfer learning stage uses knowledge from the previously learned classification task (with more data for training) and generalizes the semantic segmentation-based 2D object detection task. Finally, 2D detection results from the BEV image have been back-projected into 3D in the postprocessing stage. We verify results on two datasets: the KITTI 3D object detection dataset and the Ouster LiDAR-64 dataset, thus demonstrating that the proposed method is highly competitive in terms of mean average precision (mAP up to 70%) while still running at more than 30 frames per second (FPS).
机译:利用激光雷达点云数据的三维物体检测是自主驾驶系统感知的不可缺少的一部分。基于云的点立体物检测已经超过夜间摄像机更高的精确度更好的替代品。然而,大多数基于激光雷达的3D对象的方法中监督的方式,这意味着它们的状态的最先进的性能很大程度上依赖于大规模和良好标记的数据集工作,而这些带注释的数据集可能是昂贵的,以获得与仅在有限的情况下访问。转印学习是一个有希望的方法,以减少大型训练数据的要求,但现有的转印学习对象检测器主要用于2D物体检测而不是3D。在这项工作中,我们是代表该鸟瞰视图(BEV)的场景更有效地利用三维点云数据,并提出转移学习立体物检测基于点云语义分割。该模型最小化大规模训练数据的需求,从而减少了训练时间。首先,预处理级的原始点云数据过滤到的视图的特定字段内的BEV地图。其次,转移学习阶段使用知识从先前学习的分类任务(培训更多的数据),并概括了基于语义分割的2D物体检测任务。最后,从BEV图像的2D检测结果已被背投影到3D中的后处理阶段。我们在两个数据集验证的结果:KITTI立体物检测数据集和下台激光雷达-64的数据集,从而表明所提出的方法是在平均平均精度方面高度竞争(MAP高达70%),同时仍然在超过30运行帧每秒(FPS)。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
代理获取

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号