...
首页> 外文期刊>Information Sciences: An International Journal >Analysis of music/speech via integration of audio content and functional brain response
【24h】

Analysis of music/speech via integration of audio content and functional brain response

机译:通过整合音频内容和功能性大脑反应来分析音乐/语音

获取原文
获取原文并翻译 | 示例
   

获取外文期刊封面封底 >>

       

摘要

Effective analysis of music/speech data such as clustering, retrieval, and classification has received significant attention in recent years. Traditional methods mainly rely on the low-level acoustic features derived from digital audio stream, and the accuracy of these methods is limited by the well-known semantic gap. To alleviate this problem, we propose a novel framework for music/speech clustering, retrieval, and classification by integrating the low-level acoustic features derived from audio content with the functional magnetic resonance imaging (fMRI) measured features that represent the brain's functional response when subjects are listening to the music/speech excerpts. First, the brain networks and regions of interest (ROIs) involved in the comprehension of audio stimuli, such as the auditory, emotion, attention, and working memory systems, are located by a new approach named dense individualized and common connectivity-based cortical landmarks (DICC-COLs). Then the functional connectivity matrix measuring the similarity between the fMRI signals of different ROIs is adopted to represent the brain's comprehension of audio semantics. Afterwards, we propose an improved twin Gaussian process (ITGP) model based on self-training to predict the fMRI-measured features of testing data without fMRI scanning. Finally, multi-view learning algorithms are proposed to integrate acoustic features with fMRI-measured features for music/speech clustering, retrieval, and classification, respectively. The experimental results demonstrate the superiority of our proposed work in comparison with existing methods and suggest the advantage of integrating functional brain responses via fMRI data for music/speech analysis. (C) 2014 Elsevier Inc. All rights reserved.
机译:近年来,对音乐/语音数据(例如聚类,检索和分类)的有效分析受到了广泛的关注。传统方法主要依靠从数字音频流中导出的低级声学特征,而这些方法的准确性受到众所周知的语义鸿沟的限制。为了缓解这个问题,我们提出了一种新颖的框架,通过将音频内容中的低级声学特征与功能磁共振成像(fMRI)测得的特征相集成,从而将音乐/语音进行聚类,检索和分类,该功能代表了大脑的功能反应。受试者正在听音乐/语音摘录。首先,通过一种名为密集的个性化和基于公共连通性的皮质地标的新方法来定位涉及听觉,情感,注意力和工作记忆系统等音频刺激的大脑网络和感兴趣区域(ROI)。 (DICC-COL)。然后,采用测量不同ROI的fMRI信号之间相似度的功能连通性矩阵来表示大脑对音频语义的理解。之后,我们提出了一种基于自我训练的改进双胞胎高斯过程(ITGP)模型,无需fMRI扫描即可预测经fMRI测量的测试数据特征。最后,提出了多视图学习算法,以将声学特征与fMRI测量的特征集成在一起,分别用于音乐/语音聚类,检索和分类。实验结果表明,与现有方法相比,我们的拟议工作具有优越性,并暗示了通过功能磁共振成像数据整合功能性大脑反应进行音乐/语音分析的优势。 (C)2014 Elsevier Inc.保留所有权利。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号