首页> 外文会议>IEEE International Conference on Acoustics, Speech and Signal Processing >A deep bidirectional long short-term memory based multi-scale approach for music dynamic emotion prediction
【24h】

A deep bidirectional long short-term memory based multi-scale approach for music dynamic emotion prediction

机译:一种基于深度双向长短期记忆的多尺度音乐动态情感预测方法

获取原文

摘要

Music Dynamic Emotion Prediction is a challenging and significant task. In this paper, We adopt the dimensional valence-arousal (V-A) emotion model to represent the dynamic emotion in music. Considering the high context correlation among the music feature sequence and the advantage of Bidirectional Long Short-Term Memory (BLSTM) in capturing sequence information, we propose a multi-scale approach, Deep BLSTM (DBLSTM) based multi-scale regression and fusion with Extreme Learning Machine (ELM), to predict the V-A values in music. We achieved the best performance on the database of Emotion in Music task in MediaEval 2015 compared with other submitted results. The experimental results demonstrated the effectiveness of our novel proposed multi-scale DBLSTM-ELM model.
机译:音乐动态情感预测是一项具有挑战性且意义重大的任务。在本文中,我们采用维数价(V-A)情感模型来表示音乐中的动态情感。考虑到音乐特征序列之间的高度上下文相关性以及双向长短期记忆(BLSTM)在捕获序列信息方面的优势,我们提出了一种多尺度方法,基于Deep BLSTM(DBLSTM)的多尺度回归以及与Extreme的融合学习机(ELM),用于预测音乐中的VA值。与其他提交的结果相比,我们在MediaEval 2015中的“音乐中的情感”任务数据库上取得了最佳性能。实验结果证明了我们提出的新颖的多尺度DBLSTM-ELM模型的有效性。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号