【24h】

Multi-stream Confidence Analysis for Audio-Visual Affect Recognition

机译:视听情感识别的多流置信度分析

获取原文
获取原文并翻译 | 示例

摘要

Changes in a speaker's emotion are a fundamental component in human communication. Some emotions motivate human actions while others add deeper meaning and richness to human interactions. In this paper, we explore the development of a computing algorithm that uses audio and visual sensors to recognize a speaker's affective state. Within the framework of Multi-stream Hidden Markov Model (MHMM), we analyze audio and visual observations to detect 11 cognitive/emotive states. We investigate the use of individual modality confidence measures as a means of estimating weights when combining likelihoods in the audio-visual decision fusion. Person-independent experimental results from 20 subjects in 660 sequences suggest that the use of stream exponents estimated on training data results in classification accuracy improvement of audio-visual affect recognition.
机译:说话者情绪的变化是人类交流的基本组成部分。有些情绪可以激发人类的行动,而另一些情绪则可以为人类的互动增加更深的意义和丰富性。在本文中,我们探索了一种计算算法的开发,该算法使用音频和视觉传感器来识别说话者的情感状态。在多流隐马尔可夫模型(MHMM)的框架内,我们分析了音频和视觉观察以检测11种认知/情绪状态。当在视听决策融合中结合可能性时,我们调查使用个体模态置信度度量作为权重估计的一种方法。来自660个序列的20个受试者的独立于人的实验结果表明,使用训练数据上估计的流指数可提高视听影响识别的分类准确性。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号