首页> 外文会议>IEEE International Conference on Acoustics, Speech and Signal Processing >Piano music transcription modeling note temporal evolution
【24h】

Piano music transcription modeling note temporal evolution

机译:钢琴音乐转录建模注意时态演变

获取原文

摘要

Automatic music transcription (AMT) is the process of converting an acoustic musical signal into a symbolic musical representation such as a MIDI piano roll, which contains the pitches, the onsets and offsets of the notes and, possibly, their dynamic and source (i.e., instrument). Existing algorithms for AMT commonly identify pitches and their saliences in each frame and then form notes in a post-processing stage, which applies a combination of thresholding, pruning and smoothing operations. Very few existing methods consider the note temporal evolution over multiple frames during the pitch identification stage. In this work we propose a note-based spectrogram factorization method that uses the entire temporal evolution of piano notes as a template dictionary. The method uses an artificial neural network to detect note onsets from the audio spectral flux. Next, it estimates the notes present in each audio segment between two successive onsets with a greedy search algorithm. Finally, the spectrogram of each segment is factorized using a discrete combination of note templates comprised of full note spectrograms of individual piano notes sampled at different dynamic levels. We also propose a new psychoacoustically informed measure for spectrogram similarity.
机译:自动音乐转录(AMT)是将声学音乐信号转换为诸如MIDI钢琴卷之类的象征性音乐表示的过程,其中包含音高,音符的起音和偏移以及可能的音符和动感(例如,仪器)。现有的AMT算法通常会在每个帧中识别音高及其显着性,然后在后期处理阶段形成音符,该阶段将阈值,修剪和平滑操作组合在一起。在音高识别阶段,很少有现有方法考虑音符在多个帧上的时间演变。在这项工作中,我们提出了一种基于音符的频谱图分解方法,该方法将钢琴音符的整个时间演变用作模板字典。该方法使用人工神经网络从音频频谱通量中检测音符发作。接下来,它使用贪婪搜索算法估计两次连续发作之间每个音频片段中出现的音符。最后,使用音符模板的离散组合来分解每个片段的声谱图,该音符模板包括以不同动态水平采样的单个钢琴音符的完整音符谱图。我们还提出了一种新的心理听觉知觉的频谱图相似性度量。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号