首页> 外文期刊>IEICE Transactions on Information and Systems >Multimodal Affect Recognition Using Boltzmann Zippers
【24h】

Multimodal Affect Recognition Using Boltzmann Zippers

机译:使用玻尔兹曼拉链的多峰影响识别

获取原文
获取原文并翻译 | 示例
       

摘要

This letter presents a novel approach for automatic multi-modal affect recognition. The audio and visual channels provide complementary information for human affective states recognition, and we utilize Boltzmann zippers as model-level fusion to learn intrinsic correlations between the different modalities. We extract effective audio and visual feature streams with different time scales and feed them to two component Boltzmann chains respectively. Hidden units of the two chains are interconnected to form a Boltzmann zipper which can effectively avoid local energy minima during training. Second-order methods are applied to Boltzmann zippers to speed up learning and pruning process. Experimental results on audio-visual emotion data recorded by ourselves in Wizard of Oz scenarios and collected from the SEMAINE naturalistic database both demonstrate our approach is robust and outperforms the state-of-the-art methods.
机译:这封信提出了一种新颖的自动多模式情感识别方法。视听通道为人类的情感状态识别提供了补充信息,我们利用Boltzmann拉链作为模型级融合来学习不同模态之间的内在联系。我们提取具有不同时间尺度的有效音频和视觉特征流,并将其分别馈入两个组成的玻耳兹曼链。两条链的隐藏单元相互连接以形成玻尔兹曼拉链,可以有效避免训练过程中的局部能量最小值。将二阶方法应用于Boltzmann拉链以加快学习和修剪过程。由我们自己在《绿野仙踪》场景中记录并从SEMAINE自然数据库收集的视听情感数据的实验结果都表明,我们的方法是可靠的,并且优于最新方法。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号