首页> 外文期刊>Image and Vision Computing >SAANet: Spatial adaptive alignment network for object detection in automatic driving
【24h】

SAANet: Spatial adaptive alignment network for object detection in automatic driving

机译:SAANet:用于自动驾驶中物体检测的空间自适应对准网络

获取原文
获取原文并翻译 | 示例

摘要

images and point clouds are beneficial for object detection in a visual navigation module for autonomous driving. The spatial relationships between different objects at different times in a bimodal space can vary significantly. It is difficult to combine bimodal descriptions into a unified model to effectively detect objects in an efficient amount of time. In addition, conventional voxelization methods resolve point clouds into voxels at a global level, and often overlook local attributes of the voxels. To address these problems, we propose a novel fusionbased deep framework named SAANet. SAANet utilizes a spatial adaptive alignment (SAA) module to align point cloud features and image features, by automatically discovering the complementary information between point clouds and images. Specifically, we transform the point clouds into 3D voxels, and introduce local orientation encoding to represent the point clouds. Then, we use a sparse convolutional neural network to learn a point cloud feature. Simultaneously, a ResNet-like 2D convolutional neural network is used to extract an image feature. Next, the point cloud feature and image feature are fused by our SAA block to derive a comprehensive feature. Then, the labels and 3D boxes for objects are learned using a multi-task learning network Finally, an experimental evaluation on the KITFI benchmark demonstrates the advantages of our method in terms of average precision and inference time, as compared to previous state-of-the-art results for 3D object detection. (C) 2020 Elsevier B.V. All rights reserved.
机译:图像和点云有利于自动驾驶视觉导航模块中的对象检测。双峰空间中不同时间的不同对象之间的空间关系可能会发生很大变化。很难将双峰描述结合到一个统一的模型中,以在有效的时间内有效地检测物体。另外,常规体素化方法在全局级别将点云解析为体素,并且通常会忽略体素的局部属性。为了解决这些问题,我们提出了一种新颖的基于融合的深度框架SAANet。 SAANet通过自动发现点云和图像之间的补充信息,利用空间自适应对齐(SAA)模块来对齐点云特征和图像特征。具体来说,我们将点云转换为3D体素,并引入局部方向编码来表示点云。然后,我们使用稀疏卷积神经网络来学习点云特征。同时,使用类似ResNet的2D卷积神经网络提取图像特征。接下来,通过我们的SAA模块将点云特征和图像特征融合在一起,以得出综合特征。然后,使用多任务学习网络学习对象的标签和3D框。最后,在KITFI基准上进行的实验评估证明,与以前的状态相比,我们的方法在平均精度和推理时间方面具有优势3D对象检测的最新结果。 (C)2020 Elsevier B.V.保留所有权利。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号