首页> 外文会议>IEEE International Conference on Acoustics, Speech and Signal Processing >Lightly-supervised utterance-level emotion identification using latent topic modeling of multimodal words
【24h】

Lightly-supervised utterance-level emotion identification using latent topic modeling of multimodal words

机译:基于多模态词的潜在主题建模的轻度话语级情绪识别

获取原文

摘要

Research on multimodal emotion recognition has drawn much attention recently in diverse disciplines. With the increasing amount of multimodal data, unsupervised or semi-supervised learning has become highly desirable to automatically discover expression of emotion patterns in behavioral data. We present a novel approach for multimodal emotion learning using only a small amount of labels. Our approach is hinging on probabilistic latent semantic analysis (pLSA) that defines the latent variable as the emotion class, motivated by the conceptualization that human emotion acts as a latent control variable that regulates the external behavior manifestations, such as through speech and body gesture. In our approach, we represent the audio-visual information in an utterance as a bag of multimodal words. To exploit the interrelation between speech and gesture modalities, we propose a canonical correlation analysis (CCA) based vocabulary of multimodal words. Our approach has achieved promising experimental results. We have also demonstrated the superiority of the CCA-based multimodal words over those derived directly from the original cues.
机译:最近,多模式情感识别的研究已引起了各学科的广泛关注。随着多模态数据量的增加,非常需要无监督或半监督学习来自动发现行为数据中的情感模式表达。我们提出了一种仅使用少量标签进行多模式情感学习的新颖方法。我们的方法依赖于概率潜伏语义分析(pLSA),该潜伏语义分析将潜伏变量定义为情感类别,这是由于人类情感充当了调节外部行为表现(例如通过语音和身体手势)的潜伏控制变量的概念化所激发的。在我们的方法中,我们将视听信息表达为一袋多峰词。为了利用语音和手势模态之间的相互关系,我们提出了一种基于规范相关分析(CCA)的多模态词词汇。我们的方法取得了令人鼓舞的实验结果。我们还证明了基于CCA的多模式单词比直接源自原始提示的单词具有优越性。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号