首页> 外文期刊>IEEE Transactions on Biometrics, Behavior, and Identity Science >Discriminative Local Representation Learning for Cross-Modality Visible-Thermal Person Re-Identification
【24h】

Discriminative Local Representation Learning for Cross-Modality Visible-Thermal Person Re-Identification

机译:基于判别性局部表征学习的跨模态可见热重识别

获取原文
获取原文并翻译 | 示例
获取外文期刊封面目录资料

摘要

Visible-thermal person re-identification (VTReID) is a rising and challenging cross-modality retrieval task in intelligent video surveillance systems. Most attention architectures cannot explore the discriminative person representations for VTReID, especially in the thermal modality. In addition, the fine-grained middle-level semantic information has received much less attention in the part-based approaches for the cross-modality pedestrian retrieval task, resulting in limited generalization capability and poor representation robustness. This paper proposes a simple yet powerful discriminative local representation learning (DLRL) model to capture the robust local fine-grained feature representations and explore the rich semantic relationship between the learned part features. Specifically, an efficient contextual attention aggregation module (CAAM) is designed to strengthen the discriminative capability of the feature representations and explore the contextual cues for visible and thermal modalities. Then, an integrated middle-high feature learning (IMHF) method is introduced to capture the part-level salient representations, which handles the ambiguous modality discrepancy in both discriminative middle-level and robust high-level information. Moreover, a part-guided graph convolution module (PGCM) is constructed to mine the structural relationship among the part representations within each modality. The quantitative and qualitative experiments on the two benchmark datasets demonstrate that the proposed DLRL model significantly outperforms state-of-the-art methods and achieves rank-1/mAP accuracy of 92.77/82.05 on the RegDB dataset and 63.04/60.58 on the SYSU-MM01 dataset.
机译:可见光热重识别(VTReID)是智能视频监控系统中日益兴起且具有挑战性的跨模态检索任务。大多数注意力架构无法探索 VTReID 的判别性人表示,尤其是在热模态中。此外,在基于部分的跨模态行人检索任务中,细粒度的中间层次语义信息受到的关注较少,导致泛化能力有限,表示鲁棒性差。该文提出了一种简单而强大的判别性局部表示学习(DLRL)模型来捕获鲁棒的局部细粒度特征表示,并探索学习到的部分特征之间的丰富语义关系。具体而言,设计了一种高效的情境注意力聚合模块(CAAM),以增强特征表示的判别能力,并探索可见和热模态的情境线索。然后,引入一种集成中高特征学习(IMHF)方法来捕获部分级显著性表示,该方法处理了判别性中级和鲁棒性高级信息中的模糊模态差异。此外,该文构建了部分引导图卷积模块(PGCM)来挖掘各模态内部分表示之间的结构关系。在两个基准数据集上的定量和定性实验表明,所提出的DLRL模型明显优于现有方法,在RegDB数据集上实现了92.77%/82.05%的rank-1/mAP准确率,在SYSU-MM01数据集上达到了63.04%/60.58%的准确率。

著录项

获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号