首页> 外文会议>IEEE Workshop on Applications of Signal Processing to Audio and Acoustics >Learning emotion-based acoustic features with deep belief networks
【24h】

Learning emotion-based acoustic features with deep belief networks

机译:学习基于情感的声学功能,具有深度信仰网络

获取原文
获取外文期刊封面目录资料

摘要

The medium of music has evolved specifically for the expression of emotions, and it is natural for us to organize music in terms of its emotional associations. But while such organization is a natural process for humans, quantifying it empirically proves to be a very difficult task, and as such no dominant feature representation for music emotion recognition has yet emerged. Much of the difficulty in developing emotion-based features is the ambiguity of the ground-truth. Even using the smallest time window, opinions on the emotion are bound to vary and reflect some disagreement between listeners. In previous work, we have modeled human response labels to music in the arousal-valence (A-V) representation of affect as a time-varying, stochastic distribution. Current methods for automatic detection of emotion in music seek performance increases by combining several feature domains (e.g. loudness, timbre, harmony, rhythm). Such work has focused largely in dimensionality reduction for minor classification performance gains, but has provided little insight into the relationship between audio and emotional associations. In this new work we seek to employ regression-based deep belief networks to learn features directly from magnitude spectra. While the system is applied to the specific problem of music emotion recognition, it could be easily applied to any regression-based audio feature learning problem.
机译:音乐媒介专门演变为表达情绪,我们很自然地在其情绪协会方面组织音乐。但是,虽然这种组织是人类的自然过程,但量化它经验证明是一项非常艰巨的任务,并且由于音乐情感认可的主导特征表示尚未出现。发展情感的特征的大部分困难是地面真理的歧义。甚至使用最小的时间窗口,对情绪的意见必将有所不同,并反映听众之间的一些分歧。在以前的工作中,我们将人力响应标签建模到唤醒(A-V)表示影响作为时变的随机分布的音乐。通过组合若干特征域(例如响度,音色,和谐,节奏),在音乐中自动检测音乐情绪的现有方法增加。此类工作主要集中在很大程度上,对小分类绩效收益的维度降低,但对音频和情感协会之间的关系提供了很少的洞察力。在这项新工作中,我们寻求采用基于回归的深度信念网络,以直接从幅度谱学习特征。虽然系统适用于音乐情感识别的特定问题,但它可以很容易地应用于任何基于回归的音频特征学习问题。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号