首页> 外文会议>International Conference on Intelligent Computer Communication and Processing >Speech Emotion Recognition method using time-stretching in the Preprocessing Phase and Artificial Neural Network Classifiers
【24h】

Speech Emotion Recognition method using time-stretching in the Preprocessing Phase and Artificial Neural Network Classifiers

机译:在预处理阶段采用时间拉伸和人工神经网络分类器的语音情感识别方法

获取原文

摘要

Human emotions are playing a significant role in the understanding of human behaviour. There are multiple ways of recognizing human emotions, and one of them is through human speech. This paper aims to present an approach for designing a Speech Emotion Recognition (SER) system for an industrial training station. While assembling a product, the end user emotions can be monitored and used as a parameter for adapting the training station. The proposed method is using a phase vocoder for time-stretching and an Artificial Neural Network (ANN) for classification of five typical different emotions. As input for the ANN classifier, features like Mel Frequency Cepstral Coefficients (MFCCs), short-term energy, zero-crossing rate, pitch and the speech rate were extracted. The proposed method was evaluated on the Ryerson Audio-Visual Database of Emotion Speech and Song (RAVDESS) and shows promising results when compared to other methods such as zero-padding.
机译:人类的情感在理解人类行为中起着重要的作用。识别人类情感的方法有多种,其中一种是通过人类的语音。本文旨在提出一种用于工业培训站的语音情感识别(SER)系统的设计方法。在组装产品时,可以监控最终用户的情绪并将其用作适应训练站的参数。所提出的方法是使用相位声码器进行时间拉伸,并使用人工神经网络(ANN)对5种典型的不同情绪进行分类。作为ANN分类器的输入,提取了诸如梅尔频率倒谱系数(MFCC),短期能量,过零率,音调和语音速率等特征。该方法在Ryerson情绪语音和歌曲的视听数据库(RAVDESS)上进行了评估,与其他方法(如零填充)相比,显示出了可喜的结果。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号