...
首页> 外文期刊>International journal of imaging systems and technology >Bimodal recognition of affective states with the features inspired from human visual and auditory perception system
【24h】

Bimodal recognition of affective states with the features inspired from human visual and auditory perception system

机译:具有人类视觉和听觉感知系统启发的特征,对情绪状态进行双峰识别

获取原文
获取原文并翻译 | 示例
           

摘要

In this article, attention-based mechanism with the enhancement on biologically inspired network for emotion recognition is proposed. Existing bio-inspired models use multiscale and multiorientation architecture to gain discriminative power and to extract meticulous visual features. Prevailing HMAX model represents S2 layers by randomly selected prototype patches from training samples that increase the computational complexity and degrade the discerning ability. As eyes and mouth regions are the most powerful and reliable cues in determining facial emotions, they serve as the prototype patches for S2 layer in HMAX model. Audio code 4 book is constructed from mel-frequency cepstral coefficients, temporal and spectral features processed by principal component analysis. Audio and video data features are fused to train support vector machine classifier. The attained results on eNTERFACE, surrey audio-visual expressed emotion and acted facial expressions in the wild database datasets ascertain the efficiency of the proposed architecture for emotion recognition.
机译:在本文中,提出了一种基于注意力的机制,并增强了基于生物启发的网络进行情绪识别。现有的受生物启发的模型使用多尺度和多方位的体系结构来获得判别能力并提取精细的视觉特征。流行的HMAX模型通过从训练样本中随机选择的原型补丁来表示S2层,这会增加计算复杂性并降低识别能力。由于眼睛和嘴巴区域是确定面部表情的最有力和最可靠的线索,因此它们充当HMAX模型中S2层的原型补丁。音频代码4本书由梅尔频率倒谱系数,通过主成分分析处理的时间和频谱特征构成。音频和视频数据特征融合在一起以训练支持向量机分类器。在eNTERFACE,萨里视听表达的情感和面部表情在野生数据库数据集中获得的结果确定了所提出的情感识别架构的效率。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号