...
首页> 外文期刊>Applied Intelligence: The International Journal of Artificial Intelligence, Neural Networks, and Complex Problem-Solving Technologies >Context-aware pedestrian detection especially for small-sized instances with Deconvolution Integrated Faster RCNN (DIF R-CNN)
【24h】

Context-aware pedestrian detection especially for small-sized instances with Deconvolution Integrated Faster RCNN (DIF R-CNN)

机译:背景感知的行人检测特别适用于具有解卷积的小型实例集成了更快的RCNN(DIF R-CNN)

获取原文
获取原文并翻译 | 示例
   

获取外文期刊封面封底 >>

       

摘要

Pedestrian detection is a canonical problem in computer vision. Motivated by the observation that the major bottleneck of pedestrian detection lies on the different scales of pedestrian instances in images, our effort is focused on improving the detection rate, especially for small-sized pedestrians who are relatively far from the camera. In this paper, we introduce a novel context-aware pedestrian detection method by developing the Deconvolution Integrated Faster R-CNN (DIF R-CNN), in which we integrate a deconvolutional module to bring additional context information which is helpful to improve the detection accuracy for small-sized pedestrian instances. Furthermore, the state-of-the-art CNN-based model (Inception-ResNet) is exploited to provide a rich and discriminative hierarchy of feature representations. With these enhancements, a new synthetic feature map can be generated with a higher resolution and more semantic information. Additionally, atrous convolution is adopted to enlarge the receptive field of the synthetic feature map. Extensive evaluations on two challenging pedestrian detection datasets demonstrate the effectiveness of the proposed DIF R-CNN. Our new approach performs 12.29% better for detecting small-sized pedestrians (those below 50 pixels in bounding-box height) and 6.87% better for detecting all case pedestrians of the Caltech benchmark than the state-of-the-art method. For aerial-view small-sized pedestrian detection, our method achieve 8.9% better performance when compared to the baseline method on the Okutama human-action dataset.
机译:行人检测是计算机视觉中的规范问题。观察到人行为检测的主要瓶颈在图像中的行人实例的不同尺度上,我们的努力专注于提高检测率,特别是对于与相机相对较远的小型行人。在本文中,我们通过开发DeconVolution综合R-CNN(DIF R-CNN)来介绍一种新的背景知识的行人检测方法(DIF R-CNN),其中我们集成了解卷积模块,以带来额外的上下文信息,这有助于提高检测精度。对于小型行人实例。此外,利用最先进的基于CNN的模型(Inception-Reset)以提供特征表示的丰富和辨别性等级。利用这些增强功能,可以使用更高的分辨率和更多语义信息生成新的合成特征映射。另外,采用不受欢迎的卷积来扩大合成特征图的接受领域。对两个具有挑战性的行人检测数据集的广泛评估证明了所提出的DIF R-CNN的有效性。我们的新方法更好地执行了12.29%,用于检测小型行人(边界箱高度低于50像素),检测到CALTECH基准的所有案例行人比最先进的方法更好地进行6.87%。对于鸟瞰的小型行人检测,与Okutama人为行动数据集的基线方法相比,我们的方法达到了8.9%的性能。

著录项

相似文献

  • 外文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号