首页> 外文会议>International Conference on intelligent science and big data engineering >Learning Siamese Network with Top-Down Modulation for Visual Tracking
【24h】

Learning Siamese Network with Top-Down Modulation for Visual Tracking

机译:使用自顶向下调制的视觉跟踪学习暹罗网络

获取原文

摘要

The performance of visual object tracking depends largely on the target appearance model. Benefited from the success of CNN in feature extraction, recent studies have paid much attention to CNN representation learning and feature fusion model. However, the existing feature fusion models ignore the relation between the features of different layers. In this paper, we propose a deep feature fusion model based on the Siamese network by considering the connection between feature maps of CNN. To tackle the limitation of different feature map sizes in CNN, we propose to fuse different resolution feature maps by introducing de-convolutional layers in the offline training stage. Specifically, a top-down modulation is adopted for feature fusion. In the tracking stage, a simple matching operation between the fused feature of the examplar and search region is conducted with the learned model, which can maintain the real-time tracking speed. Experimental results show that, the proposed method obtains favorable tracking accuracy against the state-of-the-art trackers with a real-time tracking speed.
机译:视觉对象跟踪的性能很大程度上取决于目标外观模型。得益于CNN在特征提取中的成功应用,最近的研究非常关注CNN表示学习和特征融合模型。然而,现有的特征融合模型忽略了不同层的特征之间的关系。在本文中,我们考虑了CNN特征图之间的联系,提出了一种基于暹罗网络的深度特征融合模型。为了解决CNN中不同特征图大小的限制,我们建议在离线训练阶段通过引入反卷积层来融合不同分辨率的特征图。具体地,采用自上而下的调制进行特征融合。在跟踪阶段,利用学习的模型对示例融合特征与搜索区域之间的简单匹配操作,可以保持实时跟踪速度。实验结果表明,该方法相对于最新的跟踪器具有良好的跟踪精度,并且具有实时的跟踪速度。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号