首页> 外文会议>International Conference on Signal Processing >An Improved Multimodal Dimension Emotion Recognition Based on Different Fusion Methods
【24h】

An Improved Multimodal Dimension Emotion Recognition Based on Different Fusion Methods

机译:基于不同融合方法的改进的多模式尺寸情感识别

获取原文

摘要

Continuous emotion recognition is a challenging task and a key part of human-computer interaction, especially multimodal emotion recognition can effectively improve the accuracy and robustness of recognition. But there are limited emotion data sets, and it is difficult to extract emotion features. We present a multi-level segmented decision level fusion emotion recognition model to improve the performance of emotion recognition. In this paper, we predict multi-modal dimensional emotional state on AVEC2017 dataset. Our model uses Bidirectional Long Short-Term Memory (BLSTM) as multi-level segmented emotional feature learning model, and uses the SVR model as fusion model of the decision layer. The BLSTM can model different forms of emotional information in time, and can also consider the impact of previous and later emotional features on current results. The SVR model can compensate for the redundant information of emotion recognition. At the same time, we also consider annotation delay and temporal pooling in our multi-modal dimensional emotion recognition model. Our multi-modal emotion recognition model achieves significant recognition improvements and provide the robustness. Finally, we compare the baseline methods which used the same dataset, and find that the CCC performance of our method is the best on arousal, which is 0.685. Our research shows that the proposed multi-layer segmentation decision level fusion emotion recognition model is conducive to improving performance.
机译:连续情感识别是一项艰巨的任务和人机交互的重要组成部分,尤其是多模态的情感识别能有效地提高识别的准确性和鲁棒性。但情感数据集有限,难以提取情绪特征。我们提出了一种多级分段决策级融合情绪识别模型,以提高情感识别的表现。在本文中,我们预测AVEC2017数据集上的多模态维度情绪状态。我们的模型使用双向长期内存(BLSTM)作为多级分段情感特征学习模型,并使用SVR模型作为决策层的融合模型。 BLSTM可以及时模拟不同形式的情感信息,也可以考虑前后情绪特征对当前结果的影响。 SVR模型可以补偿情绪识别的冗余信息。与此同时,我们还考虑在我们的多模态维度情感识别模型中的注释延迟和时间汇集。我们的多模态情绪识别模型实现了重大识别改进并提供了稳健性。最后,我们比较了使用相同数据集的基线方法,并发现我们方法的CCC性能是唤醒的最佳状态,即0.685。我们的研究表明,所提出的多层分割决策电平融合情绪识别模型有利于提高性能。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号