...
首页> 外文期刊>International Journal of Applied Mathematics and Computer Science >COMPARISON OF SPEAKER DEPENDENT AND SPEAKER INDEPENDENT EMOTION RECOGNITION
【24h】

COMPARISON OF SPEAKER DEPENDENT AND SPEAKER INDEPENDENT EMOTION RECOGNITION

机译:演讲者依赖和演讲者独立情绪识别的比较

获取原文
获取原文并翻译 | 示例
   

获取外文期刊封面封底 >>

       

摘要

This paper describes a study of emotion recognition based on speech analysis. The introduction to the theory contains a review of emotion inventories used in various studies of emotion recognition as well as the speech corpora applied, methods of speech parametrization, and the most commonly employed classification algorithms. In the current study the EMO-DB speech corpus and three selected classifiers, the κ-Nearest Neighbor (κ-NN), the Artificial Neural Network (ANN) and Support Vector Machines (SVMs), were used in experiments. SVMs turned out to provide the best classification accuracy of 75.44% in the speaker dependent mode, that is, when speech samples from the same speaker were included in the training corpus. Various speaker dependent and speaker independent configurations were analyzed and compared. Emotion recognition in speaker dependent conditions usually yielded higher accuracy results than a similar but speaker independent configuration. The improvement was especially well observed if the base recognition ratio of a given speaker was low. Happiness and anger, as well as boredom and neutrality, proved to be the pairs of emotions most often confused.
机译:本文介绍了一种基于语音分析的情感识别研究。该理论的简介包括对各种用于情绪识别研究的情绪清单的回顾,以及所应用的语音语料库,语音参数化方法和最常用的分类算法。在当前的研究中,在实验中使用了EMO-DB语音语料库和三个选定的分类器,即κ最近邻(κ-NN),人工神经网络(ANN)和支持向量机(SVM)。事实证明,在说话者依存模式下(即,来自同一说话者的语音样本包含在训练语料库中),SVM可以提供75.44%的最佳分类精度。分析并比较了各种取决于扬声器和与扬声器无关的配置。与类似但独立于扬声器的配置相比,在与扬声器相关的条件下进行情绪识别通常会产生更高的准确性结果。如果给定说话者的基本识别率很低,则可以很好地观察到这种改进。幸福和愤怒以及无聊和中立被证明是最容易混淆的情绪对。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号