首页> 外文会议>IEEE Conference on Multimedia Information Processing and Retrieval >Self-Attentive Feature-Level Fusion for Multimodal Emotion Detection
【24h】

Self-Attentive Feature-Level Fusion for Multimodal Emotion Detection

机译:自养功能级融合,用于多模式情绪检测

获取原文

摘要

Multimodal emotion recognition is the task of detecting emotions present in user-generated multimedia content. Such resources contain complementary information in multiple modalities. A stiff challenge often faced is the complexity associated with feature-level fusion of these heterogeneous modes. In this paper, we propose a new feature-level fusion method based on self-attention mechanism. We also compare it with traditional fusion methods such as concatenation, outer-product, etc. Analyzed using textual and speech (audio) modalities, our results suggest that the proposed fusion method outperforms others in the context of utterance-level emotion recognition in videos.
机译:多模式情绪识别是检测用户生成的多媒体内容中存在的情绪的任务。这些资源包含多种方式的互补信息。经常面临的坚硬挑战是与这些异构模式的特征级融合相关的复杂性。本文提出了一种基于自我关注机制的新特征级融合方法。我们还使用传统的融合方法进行比较,例如使用文本和语音(音频)方式分析的倾斜,外部产品等,我们的结果表明,所提出的融合方法在视频中的话语级情感识别的背景下表现出其他人。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号