首页> 外文会议>IEEE International Conference on Multimedia and Expo >An End-to-End Learning Approach for Multimodal Emotion Recognition: Extracting Common and Private Information
【24h】

An End-to-End Learning Approach for Multimodal Emotion Recognition: Extracting Common and Private Information

机译:一种用于多模式情感识别的端到端学习方法:提取公共和私人信息

获取原文
获取外文期刊封面目录资料

摘要

Multimodal emotion recognition is important for facilitating efficient interaction between humans and machines. To better detect emotional states from multimodal data, we need to effectively extract both the common information that captures dependencies among different modalities, and the private information that characterizes variations in each modality. However, existing works are mostly designed to pursue either one of these objectives but not both. In our work, we propose an end-to-end learning approach to simultaneously extract the common and private information for multimodal emotion recognition. Specifically, we use a correlation loss based on Hirschfeld-Gebelein-Renyi (HGR) maximal correlation and a reconstruction loss based on autoencoders to preserve the common and private information, respectively. Experimental results on eNTERFACE'05 database and RML database demonstrate the effectiveness of our proposed approach.
机译:多模式情感识别对于促进人与机器之间的有效交互非常重要。为了更好地从多模态数据中检测情绪状态,我们需要有效地提取捕获不同模态之间依存关系的公共信息,以及表征每种模态变化的私人信息。但是,现有作品大多旨在实现这些目标之一,而不是两者兼而有之。在我们的工作中,我们提出了一种端到端的学习方法,可以同时提取用于多模式情感识别的公共和私人信息。具体来说,我们使用基于Hirschfeld-Gebelein-Renyi(HGR)最大相关性的相关损失和基于自动编码器的重构损失来分别保留公共信息和私有信息。在eNTERFACE'05数据库和RML数据库上的实验结果证明了我们提出的方法的有效性。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号