首页> 外文会议>IEEE International Conference on Acoustics, Speech and Signal Processing >Improved music feature learning with deep neural networks
【24h】

Improved music feature learning with deep neural networks

机译:用深神经网络改进音乐特色学习

获取原文

摘要

Recent advances in neural network training provide a way to efficiently learn representations from raw data. Good representations are an important requirement for Music Information Retrieval (MIR) tasks to be performed successfully. However, a major problem with neural networks is that training time becomes prohibitive for very large datasets and the learning algorithm can get stuck in local minima for very deep and wide network architectures. In this paper we examine 3 ways to improve feature learning for audio data using neural networks: 1.using Rectified Linear Units (ReLUs) instead of standard sigmoid units; 2.using a powerful regularisation technique called Dropout; 3.using Hessian-Free (HF) optimisation to improve training of sigmoid nets. We show that these methods provide significant improvements in training time and the features learnt are better than state of the art handcrafted features, with a genre classification accuracy of 83 ± 1.1% on the Tzanetakis (GTZAN) dataset. We found that the rectifier networks learnt better features than the sigmoid networks. We also demonstrate the capacity of the features to capture relevant information from audio data by applying them to genre classification on the ISMIR 2004 dataset.
机译:神经网络培训的最新进展提供了一种有效地从原始数据学习表示的方法。良好的表示是您成功执行的音乐信息检索(MIR)任务的重要要求。然而,神经网络的主要问题是,对于非常大的数据集来说,训练时间变得令人望而却步,并且学习算法可以在极限和广泛的网络架构中陷入局部最小值。在本文中,我们研究了使用神经网络改进音频数据的特征学习的方法:1。排除整流的线性单元(Relus)代替标准的乙状结单位; 2.使用一个称为辍学的强大正则化技术; 3.使用Hessian的(HF)优化,以改善围网训练。我们表明,这些方法在培训时间提供了显着的改进,并且学到的特征优于艺术的现状优于艺术功能,具有83±1.1%的Tzanetakis(GTZAN)数据集。我们发现整流网络学会了比Sigmoid网络更好的特征。我们还通过将其应用于ISMIR 2004 DataSet上的流派分类来展示功能从音频数据捕获相关信息的能力。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号