首页> 外文会议>2018 First Asian Conference on Affective Computing and Intelligent Interaction >Sequence-to-sequence Modelling for Categorical Speech Emotion Recognition Using Recurrent Neural Network
【24h】

Sequence-to-sequence Modelling for Categorical Speech Emotion Recognition Using Recurrent Neural Network

机译:基于递归神经网络的分类语音情感识别的序列化建模

获取原文
获取原文并翻译 | 示例

摘要

To model the categorical speech emotion recognition tasks in a sequential approach, the first challenge is how to transfer the categorical label for each utterance into a label sequence. To settle this, we make a hypothesis that an utterance is consisting of emotional and non-emotional segments alternatively, and these non-emotional segments correspond to silent regions, short pauses, transits between phonemes, fricative phonemes, etc. With this hypothesis, we propose to treat an utterance, 's label sequence as a chain of two kinds of states: emotional states denoting emotional frames and Nulls denoting non-emotional frames. Then, we exploit a connectionist temporal classification based recurrent neural network (CTC-RNN) to automatically label and align an utterance's emotional segments with emotional labels, while non-emotional segments with non-emotional labels. Experimental results on the IEMOCAP corpus demonstrate the effectiveness of our proposed method compared to state-of-the-art emotion recognition algorithms.
机译:为了用顺序方法对分类语音情感识别任务进行建模,第一个挑战是如何将每种话语的分类标签转换为标签序列。为了解决这个问题,我们做出一个假设,即话语由情感和非情感性片段组成,这些非情感性片段对应于静默区域,短暂的停顿,音素之间的过渡,摩擦音素等。建议将发话标签的序列视为两种状态的链:情绪状态表示情绪框架,空值表示非情绪框架。然后,我们利用基于连接器的时间分类的递归神经网络(CTC-RNN)自动标记并对齐带有情感标签的话语情感段,而将非情感段与非情感标签对齐。与最先进的情绪识别算法相比,IEMOCAP语料库上的实验结果证明了我们提出的方法的有效性。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号