首页> 外文期刊>Advanced Robotics: The International Journal of the Robotics Society of Japan >Emotional Feature Extraction Method Based on the Concentration of Phoneme Influence for Human-Robot Interaction
【24h】

Emotional Feature Extraction Method Based on the Concentration of Phoneme Influence for Human-Robot Interaction

机译:基于音素影响集中的人机交互情感特征提取方法

获取原文
获取原文并翻译 | 示例
           

摘要

Depending on the emotion of speech, the meaning of the speech or the intention of the speaker differs. Therefore, speech emotion recognition, as well as automatic speech recognition is necessary to communicate precisely between humans and robots for human-robot interaction. In this paper, a novel feature extraction method is proposed for speech emotion recognition using separation of phoneme class. In feature extraction, the signal variation caused by different sentences usually overrides the emotion variation and it lowers the performance of emotion recognition. However, as the proposed method extracts features from speech in parts that correspond to limited ranges of the center of gravity of the spectrum (CoG) and formant frequencies, the effects of phoneme variation on features are reduced. Corresponding to the range of CoG, the obstruent sounds are discriminated from sonorant sounds. Moreover, the sonorant sounds are categorized into four classes by the resonance characteristics revealed by formant frequency. The result shows that the proposed method using 30 different speakers' corpora improves emotion recognition accuracy compared with other methods by 99% significance level. Furthermore, the proposed method was applied to extract several features including prosodic and phonetic features, and was implemented on 'Mung' robots as an emotion recognizer of users.
机译:根据讲话的情感,讲话的含义或讲话者的意图会有所不同。因此,语音情感识别以及自动语音识别对于在人与机器人之间进行精确的通信以进行人机交互至关重要。本文提出了一种基于音素分类的语音情感识别特征提取方法。在特征提取中,由不同句子引起的信号变化通常会覆盖情绪变化,从而降低情绪识别的性能。但是,由于提出的方法从语音中提取与频谱重心(CoG)和共振峰频率的有限范围相对应的部分中的特征,因此减小了音素变化对特征的影响。对应于CoG的范围,将son的声音与from的声音区分开。而且,共振峰声音根据共振峰频率所表现出的共振特性分为四类。结果表明,与其他方法相比,使用30种不同说话者语料库的方法将情感识别准确度提高了99%。此外,所提出的方法被用于提取包括韵律和语音特征在内的多个特征,并在“蒙”(Mung)机器人上作为用户的情绪识别器实现。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号