首页> 外文会议>2018 IEEE Conference on Multimedia Information Processing and Retrieval >Self-Attentive Feature-Level Fusion for Multimodal Emotion Detection
【24h】

Self-Attentive Feature-Level Fusion for Multimodal Emotion Detection

机译:用于多模式情感检测的自专心特征级融合

获取原文
获取原文并翻译 | 示例

摘要

Multimodal emotion recognition is the task of detecting emotions present in user-generated multimedia content. Such resources contain complementary information in multiple modalities. A stiff challenge often faced is the complexity associated with feature-level fusion of these heterogeneous modes. In this paper, we propose a new feature-level fusion method based on self-attention mechanism. We also compare it with traditional fusion methods such as concatenation, outer-product, etc. Analyzed using textual and speech (audio) modalities, our results suggest that the proposed fusion method outperforms others in the context of utterance-level emotion recognition in videos.
机译:多模式情感识别是检测用户生成的多媒体内容中存在的情感的任务。这些资源包含多种形式的补充信息。通常面临的一个严峻挑战是与这些异构模式的特征级融合相关的复杂性。本文提出了一种基于自我注意机制的特征级融合方法。我们还将其与传统的融合方法(例如串联,外部产品等)进行比较。使用文本和语音(音频)形式进行分析,我们的结果表明,在视频的话语级情感识别的背景下,提出的融合方法优于其他融合方法。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号