...
首页> 外文期刊>Cybernetics, IEEE Transactions on >Time-Delay Neural Network for Continuous Emotional Dimension Prediction From Facial Expression Sequences
【24h】

Time-Delay Neural Network for Continuous Emotional Dimension Prediction From Facial Expression Sequences

机译:时延神经网络用于基于面部表情序列的连续情绪维度预测

获取原文
获取原文并翻译 | 示例

摘要

Automatic continuous affective state prediction from naturalistic facial expression is a very challenging research topic but very important in human–computer interaction. One of the main challenges is modeling the dynamics that characterize naturalistic expressions. In this paper, a novel two-stage automatic system is proposed to continuously predict affective dimension values from facial expression videos. In the first stage, traditional regression methods are used to classify each individual video frame, while in the second stage, a time-delay neural network (TDNN) is proposed to model the temporal relationships between consecutive predictions. The two-stage approach separates the emotional state dynamics modeling from an individual emotional state prediction step based on input features. In doing so, the temporal information used by the TDNN is not biased by the high variability between features of consecutive frames and allows the network to more easily exploit the slow changing dynamics between emotional states. The system was fully tested and evaluated on three different facial expression video datasets. Our experimental results demonstrate that the use of a two-stage approach combined with the TDNN to take into account previously classified frames significantly improves the overall performance of continuous emotional state estimation in naturalistic facial expressions. The proposed approach has won the affect recognition sub-challenge of the Third International Audio/Visual Emotion Recognition Challenge.
机译:从自然主义的面部表情自动进行连续的情感状态预测是一个非常具有挑战性的研究课题,但在人机交互中非常重要。主要挑战之一是对表征自然表达的动力学进行建模。本文提出了一种新颖的两阶段自动系统,用于从面部表情视频中连续预测情感维度值。在第一阶段,使用传统的回归方法对每个单独的视频帧进行分类,而在第二阶段,提出了时延神经网络(TDNN)来建模连续预测之间的时间关系。两阶段方法将情绪状态动力学建模与基于输入特征的单个情绪状态预测步骤分开。这样做,TDNN使用的时间信息不会受到连续帧特征之间的高可变性的偏见,并且使网络更容易利用情绪状态之间缓慢变化的动态。该系统在三个不同的面部表情视频数据集上进行了全面测试和评估。我们的实验结果表明,将两阶段方法与TDNN结合使用以考虑先前分类的帧,可以显着提高自然面部表情中连续情绪状态估计的整体性能。所提出的方法已经赢得了第三届国际视听情感识别挑战赛的情感识别子挑战。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号