首页> 外文会议>International Conference on Advanced Computational Intelligence >MMTrans-MT: A Framework for Multimodal Emotion Recognition Using Multitask Learning
【24h】

MMTrans-MT: A Framework for Multimodal Emotion Recognition Using Multitask Learning

机译:MMTRANS-MT:使用多址学习的多模式情感识别框架

获取原文

摘要

With the development of deep learning, emotion recognition tasks are more inclined to use multimodal data and adequate supervised information to improve accuracy. In this work, MMTrans-MT (Multimodal Transformer-Multitask), the framework for multimodal emotion recognition using multitask learning is proposed. It has three modules: modalities representation module, multimodal fusion module, and multitask output module. Three modalities, i.e, words, audio and video, are comprehensively utilized to carry out emotion recognition by a simple but efficient fusion model based on Transformer. As for multitask learning, the two tasks are defined as categorical emotion classification and dimensional emotion regression. Considering a potential mapping relationship between two kinds of emotion model, multitask learning is adopted to make the two tasks promote each other and improve recognition accuracy. We conduct experiments on CMU-MOSEI and IEMOCAP datasets. Comprehensive experiments show that the accuracy of recognition using multimodal information is higher than that using unimodal information. Adopting multitask learning promotes the performance of emotion recognition.
机译:随着深度学习的发展,情感识别任务更倾向于使用多模式数据和适当的监督信息来提高准确性。在这项工作中,提出了使用多址学习的MMTRANS-MT(多模式变压器 - MULTITASTASTASTAM),使用多址学习的多模式情感识别框架。它有三个模块:模块表示模块,多模态融合模块和多任务输出模块。三种方式,即单词,音频和视频被全面地利用基于变压器的简单但有效的融合模型来实现情绪识别。至于多任务学习,这两个任务被定义为分类情绪分类和维度情感回归。考虑到两种情绪模型之间的潜在映射关系,采用多任务学习来使两个任务互相促进并提高识别准确性。我们对CMU-MOSEI和IEMOCAP数据集进行实验。综合实验表明,使用多式联运信息的识别准确性高于使用单峰信息的识别的准确性。采用多任务学习促进情感认可的表现。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号