首页> 外文会议>IEEE International Conference on Acoustics, Speech and Signal Processing >Data Fusion for Audiovisual Speaker Localization: Extending Dynamic Stream Weights to the Spatial Domain
【24h】

Data Fusion for Audiovisual Speaker Localization: Extending Dynamic Stream Weights to the Spatial Domain

机译:视听扬声器本地化的数据融合:将动态流权重扩展到空间域

获取原文
获取外文期刊封面目录资料

摘要

Estimating the positions of multiple speakers can be helpful for tasks like automatic speech recognition or speaker diarization. Both applications benefit from a known speaker position when, for instance, applying beamforming or assigning unique speaker identities. Recently, several approaches utilizing acoustic signals augmented with visual data have been proposed for this task. However, both the acoustic and the visual modality may be corrupted in specific spatial regions, for instance due to poor lighting conditions or to the presence of background noise. This paper proposes a novel audiovisual data fusion framework for speaker localization by assigning individual dynamic stream weights to specific regions in the localization space. This fusion is achieved via a neural network, which combines the predictions of individual audio and video trackers based on their time- and location-dependent reliability. A performance evaluation using audiovisual recordings yields promising results, with the proposed fusion approach outperforming all baseline models.
机译:估算多个扬声器的位置可以有助于自动语音识别或扬声器日益增估等任务。例如,当应用波束成形或分配唯一的扬声器标识时,这两个应用程序都受益于已知的扬声器位置。最近,已经提出了为此任务提出了利用使用视觉数据增强的声学信号的几种方法。然而,声学和视觉模态都可以在特定的空间区域损坏,例如由于照明条件不佳或存在背景噪声。本文通过将单独的动态流权重分配给本地化空间中的特定区域来提出了一种新的视听数据融合框架。通过神经网络实现该融合,该网络基于它们的时间和位置相关的可靠性结合各个音频和视频跟踪器的预测。使用视听记录的性能评估产生了有希望的结果,提出的融合方法优于所有基线模型。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号