首页> 外文会议>Annual Conference on Information Sciences and Systems >A multimodal-corpus data collection system for cognitive acoustic scene analysis
【24h】

A multimodal-corpus data collection system for cognitive acoustic scene analysis

机译:一种用于认知声学场景分析的多模式 - 语料库数据收集系统

获取原文

摘要

We report on the design and the collection of a multi-modal data corpus for cognitive acoustic scene analysis. Sounds are generated by stationary and moving sources (people), that is by omni-directional speakers mounted on people's heads. One or two subjects walk along predetermined systematic and random paths, in synchrony and out of sync. Sound is captured in multiple microphone systems, including a four MEMS microphone directional array, two electret microphones situated in the ears of a stuffed gerbil head, and a Head Acoustics, head-shoulder unit with ICP microphones. Three micro-Doppler units operating at different frequencies were employed to capture gait and the articulatory signatures as well as location of the people in the scene. Three ground vibration sensors were recording the footsteps of the walking people. A 3D MESA camera as well as a web-cam provided 2D and 3D visual data for system calibration and ground truth. Data were collected in three environments ranging from a well controlled environment (anechoic chamber), an indoor environment (large classroom) and the natural environment of an outside courtyard. A software tool has been developed for the browsing and visualization of the data.
机译:我们报告了一种用于认知声学场景分析的多模态数据语料库的设计和集合。声音由静止和移动来源(人)产生,即由安装在人们的头上的全方位扬声器。一个或两个受试者沿着预定的系统和随机路径散步,同步和异步。声音在多个麦克风系统中捕获,包括四个MEMS麦克风方向阵列,位于填充的Gerbil头部的耳朵中的两个驻极体麦克风,以及带有ICP麦克风的头部声学,头部肩部单元。在不同频率下操作的三个微多普勒单元被用于捕获步态和铰接性签名以及场景中人员的位置。三个地面振动传感器正在录制行走人的脚步。 3D MESA相机以及Web-CaM提供2D和3D可视数据,用于系统校准和地面真理。在三个环境中收集的数据,范围从受控环境(Anechoice Champer),室内环境(大型教室)和外部庭院的自然环境。已经开发了一种软件工具,用于浏览和可视化数据。

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号