首页> 外文会议>IEEE International Symposium on Circuits and Systems >A Multi-Modal Approach to Emotion Recognition using Undirected Topic Models
【24h】

A Multi-Modal Approach to Emotion Recognition using Undirected Topic Models

机译:一种使用无向主题模型的情感识别的多模态方法

获取原文

摘要

A multi-modal framework for emotion recognition using bag-of-words features and undirected, replicated softmax topic models is proposed here. Topic models ignore the temporal information between features, allowing them to capture the complex structure without a brute-force collection of statistics. Experiments are performed over face, speech and language features extracted from the USC IEMOCAP database. Performance on facial features yields an unweighted average recall of 60.71%, a relative improvement of 8.89% over state-of-the-art approaches. A comparable performance is achieved when considering only speech (57.39%) or a fusion of speech and face information (66.05%). Individually, each source is shown to be strong at recognizing either sadness (speech) or happiness (face) or neutral (language) emotions, while, a multi-modal fusion retains these properties and improves the accuracy to 68.92%. Implementation time for each source and their combination is provided. Results show that a turn of 1 second duration can be classified in approximately 666.65ms, thus making this method highly amenable for real-time implementation.
机译:在此提出了一种使用单词袋功能和无向复制的SoftMax主题模型的情感识别的多模态框架。主题模型忽略功能之间的时间信息,允许它们捕获复杂结构而不存在统计数据集。实验在USC IEMocap数据库中提取的面部,语音和语言功能上进行。面部特征的性能产生了60.71%的未加权平均召回,相对提高了最先进的方法8.89%。在考虑语音(57.39%)或言语和面部信息的融合时,可以实现相当的性能(66.05%)。单独地,每个来源都被认为是坚强的识别悲伤(语音)或幸福(面部)或中性(语言)情绪,而多模态融合保留这些性质,并将准确性提高到68.92%。提供了每个源和它们的组合的实现时间。结果表明,转弯1秒钟可以在大约666.65ms中进行分类,从而使该方法高度适用于实时实现。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号