首页> 外文会议>International conference on cognitive computing;Services conference federation >Speech Emotion Recognition Using Multi-granularity Feature Fusion Through Auditory Cognitive Mechanism
【24h】

Speech Emotion Recognition Using Multi-granularity Feature Fusion Through Auditory Cognitive Mechanism

机译:基于听觉认知机制的多粒度特征融合语音情感识别

获取原文

摘要

In this paper, we focus on the problems of single granularity in feature extraction, loss of temporal information and inefficient use of frame features in discrete speech emotion recognition. Firstly, preliminary cognitive mechanism of auditory emotion is explored through cognitive experiments, and then a multi-granularity fusion feature extraction method inspired by the mechanism for discrete emotional speech signals is proposed. The method can extract 3 different granularity features, including short-term dynamic features of frame granularity, dynamic features of segment granularity and long-term static features of global granularity. Finally, we use the LSTM network model to classify emotions according to the long-term and short-term characteristics of the fusion features. We implement experiment on the discrete emotion datasets of CHEAVD (CASIA Chinese Emotional Audio-Visual Database) released by the Institute of automation, China Research Academy of Sciences, and achieved improvement in recognition rate, increasing the MAP by 6.48%.
机译:在本文中,我们关注于特征提取中的单个粒度,时间信息的丢失以及离散语音情感识别中帧特征使用效率低下的问题。首先通过认知实验探索听觉情绪的初步认知机制,然后提出一种基于离散情绪语音信号机制的多粒度融合特征提取方法。该方法可以提取三种不同的粒度特征,包括帧粒度的短期动态特征,段粒度的动态特征和全局粒度的长期静态特征。最后,我们使用LSTM网络模型根据融合特征的长期和短期特征对情绪进行分类。我们对由中国科学院自动化研究所发布的CHEAVD(CASIA中国情感视听数据库)的离散情感数据集进行了实验,从而提高了识别率,使MAP提高了6.48%。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号