首页> 外文会议>IEEE/CVF Conference on Computer Vision and Pattern Recognition >Spatial-Temporal Graph Convolutional Network for Video-Based Person Re-Identification
【24h】

Spatial-Temporal Graph Convolutional Network for Video-Based Person Re-Identification

机译:基于时空图卷积网络的视频人识别

获取原文

摘要

While video-based person re-identification (Re-ID) has drawn increasing attention and made great progress in recent years, it is still very challenging to effectively overcome the occlusion problem and the visual ambiguity problem for visually similar negative samples. On the other hand, we observe that different frames of a video can provide complementary information for each other, and the structural information of pedestrians can provide extra discriminative cues for appearance features. Thus, modeling the temporal relations of different frames and the spatial relations within a frame has the potential for solving the above problems. In this work, we propose a novel Spatial-Temporal Graph Convolutional Network (STGCN) to solve these problems. The STGCN includes two GCN branches, a spatial one and a temporal one. The spatial branch extracts structural information of a human body. The temporal branch mines discriminative cues from adjacent frames. By jointly optimizing these branches, our model extracts robust spatial-temporal information that is complementary with appearance information. As shown in the experiments, our model achieves state-of-the-art results on MARS and DukeMTMC-VideoReID datasets.
机译:尽管基于视频的人员重新识别(Re-ID)近年来引起了越来越多的关注并取得了长足的进步,但是要有效克服视觉相似的负样本的遮挡问题和视觉歧义问题,仍然是非常具有挑战性的。另一方面,我们观察到视频的不同帧可以为彼此提供补充信息,而行人的结构信息可以为外观特征提供额外的判别线索。因此,对不同帧的时间关系和帧内的空间关系进行建模具有解决上述问题的潜力。在这项工作中,我们提出了一种新颖的时空图卷积网络(STGCN)来解决这些问题。 STGCN包括两个GCN分支,一个空间分支和一个时间分支。空间分支提取人体的结构信息。时态分支从相邻帧中挖掘判别线索。通过共同优化这些分支,我们的模型提取了与外观信息互补的鲁棒的时空信息。如实验所示,我们的模型在MARS和DukeMTMC-VideoReID数据集上获得了最先进的结果。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号