首页> 外文期刊>Neurocomputing >Object-adaptive LSTM network for real-time visual tracking with adversarial data augmentation
【24h】

Object-adaptive LSTM network for real-time visual tracking with adversarial data augmentation

机译:自适应对象LSTM网络,通过对抗性数据增强进行实时视觉跟踪

获取原文
获取原文并翻译 | 示例
           

摘要

In recent years, deep learning based visual tracking methods have obtained great success owing to the powerful feature representation ability of Convolutional Neural Networks (CNNs). Among these methods, classification-based tracking methods exhibit excellent performance while their speeds are heavily limited by the expensive computation for massive proposal feature extraction. In contrast, matching-based tracking methods (such as Siamese networks) possess remarkable speed superiority. However, the absence of online updating renders these methods unadaptable to significant object appearance variations. In this paper, we propose a novel real-time visual tracking method, which adopts an object-adaptive LSTM network to effectively capture the video sequential dependencies and adaptively learn the object appearance variations. For high computational efficiency, we also present a fast proposal selection strategy, which utilizes the matching-based tracking method to pre-estimate dense proposals and selects high-quality ones to feed to the LSTM network for classification. This strategy efficiently filters out some irrelevant proposals and avoids the redundant computation for feature extraction, which enables our method to operate faster than conventional classification-based tracking methods. In addition, to handle the problems of sample inadequacy and class imbalance during online tracking, we adopt a data augmentation technique based on the Generative Adversarial Network (GAN) to facilitate the training of the LSTM network. Extensive experiments on four visual tracking benchmarks demonstrate the state-of-the-art performance of our method in terms of both tracking accuracy and speed, which exhibits great potentials of recurrent structures for visual tracking. (C) 2019 Elsevier B.V. All rights reserved.
机译:近年来,基于深度学习的视觉跟踪方法由于卷积神经网络(CNN)强大的特征表示能力而获得了巨大的成功。在这些方法中,基于分类的跟踪方法表现出出色的性能,而其速度却受到大规模提案特征提取的昂贵计算的严重限制。相反,基于匹配的跟踪方法(例如Siamese网络)具有明显的速度优势。但是,由于缺少在线更新,因此这些方法不适用于明显的对象外观变化。在本文中,我们提出了一种新颖的实时视觉跟踪方法,该方法采用对象自适应LSTM网络来有效地捕获视频序列依存关系并自适应地学习对象外观变化。为了提高计算效率,我们还提出了一种快速的提案选择策略,该策略利用基于匹配的跟踪方法来预先估计密集的提案,然后选择高质量的提案以供LSTM网络分类。这种策略有效地过滤掉了一些不相关的建议,并避免了多余的特征提取计算,这使我们的方法比传统的基于分类的跟踪方法运行得更快。另外,为了处理在线跟踪过程中样本不足和类别不平衡的问题,我们采用基于对抗性生成网络(GAN)的数据增强技术来促进LSTM网络的训练。在四个视觉跟踪基准上进行的大量实验证明了我们的方法在跟踪准确性和速度方面都具有最先进的性能,这为视觉跟踪展现了循环结构的巨大潜力。 (C)2019 Elsevier B.V.保留所有权利。

著录项

  • 来源
    《Neurocomputing》 |2020年第7期|67-83|共17页
  • 作者

  • 作者单位

    Xiamen Univ Sch Informat Xiamen 361005 Fujian Peoples R China|Tsinghua Univ Inst Interdisciplinary Informat Sci Beijing 100084 Peoples R China;

    Xiamen Univ Sch Informat Xiamen 361005 Fujian Peoples R China;

    Xiamen Univ Technol Sch Comp & Informat Engn Xiamen 361024 Fujian Peoples R China;

    Queens Univ Belfast Sch Elect Elect Engn & Comp Sci Belfast Antrim North Ireland;

  • 收录信息 美国《科学引文索引》(SCI);美国《工程索引》(EI);
  • 原文格式 PDF
  • 正文语种 eng
  • 中图分类
  • 关键词

    Visual tracking; LSTM network; Generative adversarial network; Data augmentation;

    机译:视觉跟踪;LSTM网络;生成对抗网络;资料扩充;

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号