首页> 外文会议>IEEE International Conference on Multimedia and Expo >Learning a real-time generic tracker using convolutional neural networks
【24h】

Learning a real-time generic tracker using convolutional neural networks

机译:使用卷积神经网络学习实时通用跟踪器

获取原文

摘要

This paper presents a novel frame-pair based method for visual object tracking. Instead of adopting two-stream Convolutional Neural Networks (CNNs) to represent each frame, we stack frame pairs as the input, resulting in a single-stream CNN tracker with much fewer parameters. The proposed tracker can learn generic motion patterns of objects with much less annotated videos than previous methods. Besides, it is found that trackers trained using two successive frames tend to predict the centers of searching windows as the locations of tracked targets. To alleviate this problem, we propose a novel sampling strategy for off-line training. Specifically, we construct a pair by sampling two frames with a random offset. The offset controls the moving smoothness of objects. Experiments on the challenging VOT14 and OTB datasets show that the proposed tracker performs on par with recently developed generic trackers, but with much less memory. In addition, our tracker can run in a speed of over 100 (30) fps with a GPU (CPU), much faster than most deep neural network based trackers.
机译:本文提出了一种新颖的基于帧对的视觉目标跟踪方法。我们没有采用两流卷积神经网络(CNN)来表示每个帧,而是将帧对作为输入堆叠,从而产生了具有更少参数的单流CNN跟踪器。所提出的跟踪器可以用比以前的方法少得多的带注释的视频来学习对象的一般运动模式。此外,发现使用两个连续帧训练的跟踪器趋向于将搜索窗口的中心预测为被跟踪目标的位置。为了缓解这个问题,我们提出了一种新颖的离线训练抽样策略。具体来说,我们通过对两个具有随机偏移量的帧进行采样来构造一对。偏移量控制对象的移动平滑度。在具有挑战性的VOT14和OTB数据集上进行的实验表明,所提出的跟踪器的性能与最近开发的通用跟踪器相当,但内存却少得多。此外,我们的跟踪器通过GPU(CPU)可以以超过100(30)fps的速度运行,比大多数基于深度神经网络的跟踪器要快得多。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号