首页> 外文期刊>Journal of visual communication & image representation >Video facial emotion recognition based on local enhanced motion history image and CNN-CTSLSTM networks
【24h】

Video facial emotion recognition based on local enhanced motion history image and CNN-CTSLSTM networks

机译:基于局部增强运动历史图像和CNN-CTSLSTM网络的视频面部情感识别

获取原文
获取原文并翻译 | 示例
       

摘要

This paper focuses on the issue of recognition of facial emotion expressions in video sequences and proposes an integrated framework of two networks: a local network, and a global network, which are based on local enhanced motion history image (LEMHI) and CNN-LSTM cascaded networks respectively. In the local network, frames from unrecognized video are aggregated into a single frame by a novel method, LEMHI. This approach improves MHI by using detected human facial landmarks as attention areas to boost local value in difference image calculation, so that the action of crucial facial unit can be captured effectively. Then this single frame will be fed into a CNN network for prediction. On the other hand, an improved CNN-LSTM model is used as a global feature extractor and classifier for video facial emotion recognition in the global network. Finally, a random search weighted summation strategy is conducted as late-fusion fashion to final predication. Our work also offers an insight into networks and visible feature maps from each layer of CNN to decipher which portions of the face influence the networks' predictions. Experiments on the AFEW, CK+ and MMI datasets using subject-independent validation scheme demonstrate that the integrated framework of two networks achieves a better performance than using individual network separately. Compared with state-of-the-arts methods, the proposed framework demonstrates a superior performance. (C) 2018 Published by Elsevier Inc.
机译:本文着重于视频序列中面部表情的识别问题,并提出了两个网络的集成框架:一个局域网和一个全球网络,它们基于本地增强运动历史图像(LEMHI)和CNN-LSTM级联网络。在局域网中,无法识别的视频的帧通过一种新方法LEMHI聚合为单个帧。该方法通过将检测到的人脸标志物作为关注区域来提高差值图像计算中的局部值,从而提高了三菱重工(MHI),从而可以有效地捕获关键脸部单元的动作。然后,该单个帧将被馈入CNN网络以进行预测。另一方面,改进的CNN-LSTM模型用作全局网络中视频面部情感识别的全局特征提取器和分类器。最后,采用随机搜索加权求和策略,以后期融合的方式进行最终预测。我们的工作还提供了对CNN每一层的网络和可见特征图的深入了解,以破译面部的哪些部分会影响网络的预测。使用与主题无关的验证方案对AFEW,CK +和MMI数据集进行的实验表明,与单独使用单个网络相比,两个网络的集成框架可实现更好的性能。与最新技术方法相比,所提出的框架表现出了卓越的性能。 (C)2018由Elsevier Inc.发布

著录项

  • 来源
  • 作者单位

    Hefei Univ Technol, Sch Comp & Informat, Hefei, Anhui, Peoples R China|Anhui Prov Key Lab Affect Comp & Adv Intelligent, Hefei 230009, Anhui, Peoples R China;

    Hefei Univ Technol, Sch Comp & Informat, Hefei, Anhui, Peoples R China|Anhui Prov Key Lab Affect Comp & Adv Intelligent, Hefei 230009, Anhui, Peoples R China;

    Hefei Univ Technol, Sch Comp & Informat, Hefei, Anhui, Peoples R China|Anhui Prov Key Lab Affect Comp & Adv Intelligent, Hefei 230009, Anhui, Peoples R China;

    Hefei Univ Technol, Sch Comp & Informat, Hefei, Anhui, Peoples R China;

    Hefei Univ Technol, Sch Comp & Informat, Hefei, Anhui, Peoples R China;

  • 收录信息
  • 原文格式 PDF
  • 正文语种 eng
  • 中图分类
  • 关键词

    Video emotion recognition; Motion history image; LSTM; Facial landmarks;

    机译:视频情感识别;运动历史图像;LSTM;面部标志;
  • 入库时间 2022-08-18 04:20:24

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号