【24h】

Features extraction for speech emotion

机译:语音情感特征提取

获取原文
获取原文并翻译 | 示例
获取外文期刊封面目录资料

摘要

In this paper the speech emotion verification using two most popular methods in speech processing and analysis based on the Mel-Frequency Cepstral Coefficient (MFCC) and the Gaussian Mixture Model (GMM) were proposed and analyzed. In both cases, eatures for the speech emotion were extracted using the Short Time Fourier Transform (STFT) and Short Time Histogram (STH) for MFCC and GMM respectively. The performance of the speech emotion verification is measured based on three neural network (NN) and fuzzy neural network (FNN) architectures; namely: Multi Layer Perceptron (MLP), Adaptive Neuro Fuzzy Inference System (ANFIS) and Generic Self-organizing Fuzzy Neural Network (GenSoFNN). Results obtained from the experiments using real audio clips from movies and television sitcoms show the potential of using the proposed features extraction methods for real time application due to its reasonable accuracy and fast training time. This may lead us to the practical usage if the emotion verifier can be embedded in real time applications especially for personal digital assistance (PDA) or smart-phones.
机译:提出并分析了基于梅尔频谱倒谱系数(MFCC)和高斯混合模型(GMM)的两种语音处理和分析中最流行的语音情感验证方法。在这两种情况下,分别使用MFCC和GMM的短时傅立叶变换(STFT)和短时直方图(STH)提取语音情感特征。语音情感验证的性能基于三种神经网络(NN)和模糊神经网络(FNN)架构进行测量;分别是:多层感知器(MLP),自适应神经模糊推理系统(ANFIS)和通用自组织模糊神经网络(GenSoFNN)。使用来自电影和电视情景喜剧的真实音频片段进行的实验获得的结果表明,由于其合理的准确性和快速的训练时间,可以将建议的特征提取方法用于实时应用。如果可以将情感验证器嵌入到实时应用程序中,尤其是个人数字助理(PDA)或智能手机,那么这可能会带给我们实际使用。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号