首页> 外文学位 >Cooperative music retrieval based on automatic indexing of music by instruments and their types.
【24h】

Cooperative music retrieval based on automatic indexing of music by instruments and their types.

机译:基于乐器及其类型的自动索引音乐的合作音乐检索。

获取原文
获取原文并翻译 | 示例

摘要

With the fast booming of online music repositories, there are increasing needs for content-based Automatic Indexing to help users find their favorite music objects. Music instrument recognition is one of the main subtasks. Recently, numerous successful approaches on musical data feature extraction and selection have been proposed for instrument recognition in monophonic sounds. Unfortunately, none of those algorithms can be successfully applied to polyphonic sounds. Thus, identification of music instruments in polyphonic sounds is still difficult and challenging, especially when harmonic partials are overlapping with each other. This has stimulated the research on music sound separation and new features development for content-based automatic music information retrieval. Based on recent research results in sound classification of monophonic sounds and studies in speech recognition, Moving Picture Experts Group (MPEG) standardized a set of features of the digital audio content data for the purpose of interpretation of the information meaning for audio signal. Most of them are in a form of large matrix or a vector of large size, which are not suitable for traditional data mining algorithms; while other features in a smaller size are not sufficient for instrument recognition in polyphonic sounds. Therefore, these acoustical features themselves alone cannot be successfully applied to classification of polyphonic sounds. However, these features contain critical information, which implies music instruments' signatures.; The ultimate goal of this thesis is to build a flexible query answering system, for a musical database, retrieving from it all objects satisfying queries like "find all musical pieces in pentatonic scale with a viola and piano where viola is playing for minimum 20 seconds and piano for minimum 10 seconds". To achieve that, first of all a database of sounds containing musical instruments allowed in queries has to be built. This database is already built as a part of the music information retrieval system, called MIRAI, and it already contains about 4000 sounds taken from the MUMs (McGill University Master Samples). The descriptions of these sounds are in terms of standard musical features which definitions are provided by MPEG7, in terms of other features used earlier in a similar research, and new features proposed in this thesis. All these features are implemented and tested for their correctness. The database of musical sounds is used as a vehicle to construct several classifiers for automatic instrument recognition. In this thesis we limit our investigations to classifiers provided by WEKA and RSES (Rough Sets Exploration System). Their performance is compared against the performance of similar classifiers constructed from the same database projected to MPEG7 type features only. The main problem facing this thesis is not only the construction of the proper and sufficient set of features needed to represent musical sounds which guarantees that their descriptions can differentiate them but also a mechanism of splitting multiple instruments played simultaneously in musical sounds. For checking the performance of classifiers 3-cross or/and 10-cross validation, and bootstrap procedures are used. The classifiers showing the best performance are adopted for automatic indexing of musical pieces by instruments. Each musical piece is seen as a segmented object in the musical database with segments showing when each relevant instrument starts and ends playing. This way the musical database can be represented as an FS-tree (Frame Segment Tree) structure. The query answering system should be seen as the interface to the FS-tree representation of the musical database. The flexibility of the query answering system is based on the hierarchical structure representing all musical instruments. When a query fails, it is generalized and checked for success by the query answering system. The construction of the Flexible Query Answ
机译:随着在线音乐存储库的快速发展,对基于内容的自动索引的需求日益增长,以帮助用户找到他们喜欢的音乐对象。乐器识别是主要的子任务之一。最近,已经提出了许多关于音乐数据特征提取和选择的成功方法,用于单音中的乐器识别。不幸的是,这些算法都不能成功地应用于和弦声音。因此,以复音形式识别乐器仍然是困难且具有挑战性的,尤其是当谐波部分彼此重叠时。这刺激了对音乐声音分离的研究以及基于内容的自动音乐信息检索的新功能开发。基于单声道声音的声音分类的最新研究结果以及语音识别的研究,运动图像专家组(MPEG)标准化了数字音频内容数据的一组功能,以解释音频信号的信息含义。它们中的大多数以大矩阵或大向量的形式出现,不适合传统的数据挖掘算法;而其他较小尺寸的功能不足以识别和弦声音中的乐器。因此,这些声学特征本身不能单独成功地用于复音的分类。但是,这些功能包含关键信息,这暗示着乐器的签名。本论文的最终目标是为音乐数据库建立一个灵活的查询应答系统,从中检索所有满足查询条件的对象,例如“用中提琴和钢琴以五分音阶查找所有音乐作品,其中中提琴演奏至少20秒,然后钢琴至少10秒钟”。为此,首先必须建立一个包含查询允许的乐器的声音数据库。该数据库已经作为名为MIRAI的音乐信息检索系统的一部分而建立,并且已经包含从MUM(麦吉尔大学主样本)中提取的大约4000种声音。这些声音的描述是根据标准音乐特征(由MPEG7提供的定义),根据类似研究中较早使用的其他特征以及本文提出的新特征来进行的。所有这些功能均已实现并经过了正确性测试。音乐声数据库用作构建自动仪器识别的几个分类器的工具。在本文中,我们将研究限于WEKA和RSES(粗糙集探索系统)提供的分类器。将它们的性能与仅从投影到MPEG7类型特征的相同数据库构造的类似分类器的性能进行比较。本论文面临的主要问题不仅是表示音乐声音所需的适当而足够的一组特征的构造(保证它们的描述可以区分它们),而且是一种拆分多个同时演奏的乐器的机制。为了检查分类器的性能,使用了3交叉或/和10交叉验证以及引导程序。使用表现最佳性能的分类器来通过乐器自动索引音乐作品。每个音乐作品被视为音乐数据库中的分段对象,其中的片段显示了每个相关乐器何时开始和结束演奏。这样,音乐数据库可以表示为FS-tree(帧片段树)结构。查询应答系统应被视为音乐数据库的FS树表示的接口。查询应答系统的灵活性基于代表所有乐器的层次结构。当查询失败时,查询应答系统将其概括并检查是否成功。灵活查询答案的构造

著录项

  • 作者

    Zhang, Xin.;

  • 作者单位

    The University of North Carolina at Charlotte.$bInformation Technology (PhD).;

  • 授予单位 The University of North Carolina at Charlotte.$bInformation Technology (PhD).;
  • 学科 Information Science.; Computer Science.
  • 学位 Ph.D.
  • 年度 2007
  • 页码 114 p.
  • 总页数 114
  • 原文格式 PDF
  • 正文语种 eng
  • 中图分类 信息与知识传播;自动化技术、计算机技术;
  • 关键词

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号