首页> 外文期刊>Pattern Recognition: The Journal of the Pattern Recognition Society >Adaptive appearance model tracking for still-to-video face recognition
【24h】

Adaptive appearance model tracking for still-to-video face recognition

机译:自适应外观模型跟踪,用于从静止图像到视频的面部识别

获取原文
获取原文并翻译 | 示例
           

摘要

Systems for still-to-video face recognition (FR) seek to detect the presence of target individuals based on reference facial still images or mug-shots. These systems encounter several challenges in video surveillance applications due to variations in capture conditions (e.g., pose, scale, illumination, blur and expression) and to camera inter-operability. Beyond these issues, few reference stills are available during enrollment to design representative facial models of target individuals. Systems for still-to-video FR must therefore rely on adaptation, multiple face representation, or synthetic generation of reference stills to enhance the intra-class variability of face models. Moreover, many FR systems only match high quality faces captured in video, which further reduces the probability of detecting target individuals. Instead of matching faces captured through segmentation to reference stills, this paper exploits Adaptive Appearance Model Tracking (AAMT) to gradually learn a track-face-model for each individual appearing in the scene. The Sequential Karhunen-Loeve technique is used for online learning of these track-face-models within a particle filter-based face tracker. Meanwhile, these models are matched over successive frames against the reference still images of each target individual enrolled to the system, and then matching scores are accumulated over several frames for robust spatiotemporal recognition. A target individual is recognized if scores accumulated for a trackface-model over a fixed time surpass some decision threshold. The main advantage of AAMT over traditional still-to-video FR systems is the greater diversity of facial representation that may be captured during operations, and this can lead to better discrimination for spatiotemporal recognition. Compared to state-of-the-art adaptive biometric systems, the proposed method selects facial captures to update an individual's face model more reliably because it relies on information from tracking. Simulation results obtained with the Chokepoint video dataset indicate that the proposed method provides a significantly higher level of performance compared state-of-the-art systems when a single reference still per individual is available for matching. This higher level of performance is achieved when the diverse facial appearances that are captured in video through AAMT correspond to that of reference stills. (C) 2015 Elsevier Ltd. All rights reserved.
机译:静止图像到视频的面部识别(FR)系统试图根据参考面部静止图像或面部照片检测目标个人的存在。由于捕获条件(例如,姿势,比例,照明,模糊和表情)的变化以及摄像机的互操作性,这些系统在视频监控应用中遇到了一些挑战。除了这些问题,在注册过程中几乎没有参考静止图像可用来设计目标个人的代表性面部模型。因此,用于静态视频FR的系统必须依赖于自适应,多重面部表示或参考静态图像的合成生成,以增强面部模型的类内可变性。此外,许多FR系统仅匹配视频中捕获的高质量面部,这进一步降低了检测目标个人的可能性。本文没有将通过分割捕获的面部与参考静止图像进行匹配,而是利用自适应外观模型跟踪(AAMT)逐步学习了场景中出现的每个人的面部模型。序列Karhunen-Loeve技术用于在线学习基于粒子过滤器的人脸跟踪器中的这些人脸模型。同时,将这些模型在连续帧上与注册到系统的每个目标个人的参考静止图像进行匹配,然后在几个帧上累积匹配分数,以进行鲁棒的时空识别。如果在固定时间内累积的轨迹模型的分数超过某个决策阈值,则将识别目标个人。与传统的静态视频FR系统相比,AAMT的主要优势在于在操作过程中可能会捕获更多的面部表情,这可以更好地区分时空识别。与最新的自适应生物识别系统相比,该方法选择面部捕捉来更可靠地更新个人的面部模型,因为它依赖于跟踪信息。使用Chokepoint视频数据集获得的仿真结果表明,与每个先进的系统相比,当每个人仍然可以使用单个参考进行匹配时,所提出的方法可以提供更高水平的性能。当通过AAMT在视频中捕获的多样化面部外观与参考静止图像的面部外观相对应时,可以达到更高的性能水平。 (C)2015 Elsevier Ltd.保留所有权利。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号