首页> 外文会议>IEEE Conference on Computer Vision and Pattern Recognition Workshops >Deep Spatial-Temporal Fusion Network for Video-Based Person Re-Identification
【24h】

Deep Spatial-Temporal Fusion Network for Video-Based Person Re-Identification

机译:基于视频的人的深空间融合网络重新识别

获取原文

摘要

In this paper, we propose a novel deep end-to-end network to automatically learn the spatial-temporal fusion features for video-based person re-identification. Specifically, the proposed network consists of CNN and RNN to jointly learn both the spatial and the temporal features of input image sequences. The network is optimized by utilizing the siamese and softmax losses simultaneously to pull the instances of the same person closer and push the instances of different persons apart. Our network is trained on fullbody and part-body image sequences respectively to learn complementary representations from holistic and local perspectives. By combining them together, we obtain more discriminative features that are beneficial to person reidentification. Experiments conducted on the PRID-2011, i-LIDS-VIS and MARS datasets show that the proposed method performs favorably against existing approaches.
机译:在本文中,我们提出了一种新的深度端到端网络,可以自动学习基于视频的人的空间融合功能重新识别。具体地,所提出的网络由CNN和RNN组成,以共同学习输入图像序列的空间和时间特征。通过同时利用暹罗和SoftMax损失来缩小同一个人的实例并将不同人的实例分开,通过暹罗和Softmax损失进行优化。我们的网络分别在全体和部分身体图像序列上培训,以学习来自整体和本地观点的互补表示。通过将它们组合在一起,我们获得了对人的歧视性的特征,对人的重新入住。在Prid-2011上进行的实验,I-Lids-Vis和Mars数据集表明,该方法对现有方法有利地表现出有利。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号