首页> 外文期刊>Neurocomputing >A deep learning framework for heart rate estimation from facial videos
【24h】

A deep learning framework for heart rate estimation from facial videos

机译:面部视频心率估算的深度学习框架

获取原文
获取原文并翻译 | 示例
           

摘要

Accurate heart rate is vital to acquiring critical physical data of human subjects. For this reason, facial video-based heart rate estimation has recently received tremendous attention owing to its simplicity and convenience. However, its accuracy, reliability and computational complexity have yet to reach the desired standards. In this work, we have endeavored to develop a novel deep learning framework for real-time estimation of heart rates by using an RGB camera. Our approach consists of the following four steps. We begin Step 1 by detecting the face and facial landmarks in the video to identify the required facial Region of Interests (ROIs). In Step 2, we extract the sequence of the mean of the green-channeled video from the facial ROIs, and explore a three-stage sequential filtering, including illumination rectification, trend removal and signal amplification. In Step 3, the Short-Time Fourier Transform (STFT) is employed to convert the 1D filtered signal into the corresponding 2D Time-Frequency Representation (TFR) for characterizing the frequencies over short time intervals. The 2D TFR allows the formulation of the heart rate estimation as a video-based supervised learning problem, which can be solved by exploring a deep Convolutional Neural Network (CNN), as is carried out in Step 4. Our approach is one of the pioneering work that proposes a deep learning framework with TFRs as input for solving the heart rate estimation from facial videos. In addition, we have developed a heart rate database, named the Pulse From Face (PFF), and used it along with the existing PURE (PUlse RatE) database [1] to train the CNN. The PFF database is released for research purpose with this paper. We have evaluated the proposed framework on the MAHNOB-HCI database [2] and the VIPL-HR database [3] and compared its performance with that of other contemporary approaches to demonstrate its efficacy. (C) 2020 Published by Elsevier B.V.
机译:准确的心率对于获取人类受试者的关键物理数据至关重要。因此,由于其简单和便利性,基于面部基于视频的心率估计最近受到了巨大的关注。但是,其准确性,可靠性和计算复杂性尚未达到期望的标准。在这项工作中,我们努力通过使用RGB相机开发一个新的深度学习框架,用于使用RGB相机进行心率的实时估计。我们的方法包括以下四个步骤。我们通过检测视频中的面部和面部地标开始步骤1来识别所需的感兴趣的面部区域(ROI)。在步骤2中,我们从面部ROI中提取绿色通道视频的平均序列,并探索三级顺序滤波,包括照明整流,趋势去除和信号放大。在步骤3中,采用短时傅里叶变换(STFT)将1D滤波信号转换为相应的2D时频表示(TFR),用于在短时间间隔上表征频率。 2D TFR允许将心率估计的配方作为基于视频的监督学习问题,这可以通过探索深卷积神经网络(CNN)来解决,如步骤4所示。我们的方法是开创性之一提出与TFRS的深度学习框架的工作,作为解决面部视频的心率估计的输入。此外,我们开发了一个心率数据库,从面部(PFF)命名,并与现有的纯(脉冲率)数据库[​​1]一起使用,以培训CNN。 PFF数据库与本文有关研究目的。我们已经在Mahnob-HCI数据库[2]和Vipl-HR数据库[3]上评估了所提出的框架,并将其性能与其他当代方法的性能进行了比较,以证明其疗效。 (c)2020由elsevier b.v发布。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号