首页> 外文期刊>Neurocomputing >Fusing two-stream convolutional neural networks for RGB-T object tracking
【24h】

Fusing two-stream convolutional neural networks for RGB-T object tracking

机译:融合两流卷积神经网络进行RGB-T目标跟踪

获取原文
获取原文并翻译 | 示例
       

摘要

This paper investigates how to integrate the complementary information from RGB and thermal (RGB-T) sources for object tracking. We propose a novel Convolutional Neural Network (ConvNet) architecture, including a two-stream ConvNet and a FusionNet, to achieve adaptive fusion of different source data for robust RGB-T tracking. Both RGB and thermal streams extract generic semantic information of the target object. In particular, the thermal stream is pre-trained on the ImageNet dataset to encode rich semantic information, and then fine-tuned using thermal images to capture the specific properties of thermal information. For adaptive fusion of different modalities while avoiding redundant noises, the FusionNet is employed to select most discriminative feature maps from the outputs of the two-stream ConvNet, and updated online to adapt to appearance variations of the target object. Finally, the object locations are efficiently predicted by applying the multi-channel correlation filter on the fused feature maps. Extensive experiments on the recently public benchmark GTOT verify the effectiveness of the proposed approach against other state-of-the-art RGB-T trackers. (c) 2017 Published by Elsevier B.V.
机译:本文研究如何集成来自RGB和热(RGB-T)源的互补信息以进行对象跟踪。我们提出了一种新颖的卷积神经网络(ConvNet)架构,包括两流ConvNet和FusionNet,以实现不同源数据的自适应融合,以实现可靠的RGB-T跟踪。 RGB流和热流都提取目标对象的一般语义信息。特别是,热流在ImageNet数据集上经过预训练以编码丰富的语义信息,然后使用热图像进行微调以捕获热信息的特定属性。为了适应不同模态的自适应融合,同时避免了冗余噪声,FusionNet用于从两流ConvNet的输出中选择最具区别性的特征图,并进行在线更新以适应目标对象的外观变化。最后,通过在融合特征图上应用多通道相关滤波器,可以有效地预测目标位置。在最近公开的基准GTOT上进行的大量实验证明了该方法相对于其他最新RGB-T跟踪器的有效性。 (c)2017年由Elsevier B.V.

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号