首页> 外文会议>International conference on image analysis and recognition >Video-Based Person Re-identification by 3D Convolutional Neural Networks and Improved Parameter Learning
【24h】

Video-Based Person Re-identification by 3D Convolutional Neural Networks and Improved Parameter Learning

机译:通过3D卷积神经网络进行基于视频的人员重新识别和改进的参数学习

获取原文

摘要

In this paper we propose a novel approach for video-based person re-identification that exploits convolutional neural networks to learn the similarity of persons observed from video camera. We take 3-dimensional convolutional neural networks (3D CNN) to extract finegrained spatiotemporal features from the video sequence of a person. Unlike recurrent neural networks, 3D CNN preserves the spatial patterns of the input, which works well on re-identification problem. The network maps each video sequence of a person to a Euclidean space where distances between feature embeddings directly correspond to measures of person similarity. By our improved parameter learning method called entire triplet loss, all possible triplets in the mini-batch are taken into account to update network parameters. This parameter updating method significantly improves training, enabling the embeddings to be more discriminative. Experimental results show that our model achieves new state of the art identification rate on iLIDS-VID dataset and PRID-2011 dataset with 82.0%, 83.3% at rank 1, respectively.
机译:在本文中,我们提出了一种基于视频的人员重新识别的新方法,该方法利用卷积神经网络来学习从摄像机观察到的人员的相似性。我们采用3维卷积神经网络(3D CNN)从一个人的视频序列中提取细粒度的时空特征。与递归神经网络不同,3D CNN保留输入的空间模式,这在重新识别问题上效果很好。该网络将一个人的每个视频序列映射到一个欧几里得空间,在该空间中特征嵌入之间的距离直接对应于人的相似性度量。通过我们称为全三元组损失的改进的参数学习方法,可以考虑将迷你批处理中的所有可能的三元组考虑在内,以更新网络参数。此参数更新方法显着改善了训练,使嵌入更具区分性。实验结果表明,我们的模型在iLIDS-VID数据集和PRID-2011数据集上的识别率达到了新的水平,分别达到12.0%和83.3%。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号