...
首页> 外文期刊>Audio, Speech, and Language Processing, IEEE Transactions on >An Unsupervised Approach to Cochannel Speech Separation
【24h】

An Unsupervised Approach to Cochannel Speech Separation

机译:同道语音分离的无监督方法

获取原文
获取原文并翻译 | 示例

摘要

Cochannel (two-talker) speech separation is predominantly addressed using pretrained speaker dependent models. In this paper, we propose an unsupervised approach to separating cochannel speech. Our approach follows the two main stages of computational auditory scene analysis: segmentation and grouping. For voiced speech segregation, the proposed system utilizes a tandem algorithm for simultaneous grouping and then unsupervised clustering for sequential grouping. The clustering is performed by a search to maximize the ratio of between- and within-group speaker distances while penalizing within-group concurrent pitches. To segregate unvoiced speech, we first produce unvoiced speech segments based on onset/offset analysis. The segments are grouped using the complementary binary masks of segregated voiced speech. Despite its simplicity, our approach produces significant SNR improvements across a range of input SNR. The proposed system yields competitive performance in comparison to other speaker-independent and model-based methods.
机译:同频道(两个通话者)的语音分离主要使用预先训练的说话者相关模型来解决。在本文中,我们提出了一种用于分离同频道语音的无监督方法。我们的方法遵循计算听觉场景分析的两个主要阶段:分割和分组。对于有声语音隔离,提出的系统利用串联算法进行同时分组,然后使用无监督的分组进行顺序分组。通过搜索执行聚类,以最大化组内和组内说话者距离之比,同时惩罚组内并发音调。为了区分清音语音,我们首先根据开始/偏移分析生成清音语音段。使用隔离的有声语音的互补二进制掩码对片段进行分组。尽管它很简单,但我们的方法在整个输入SNR范围内都能显着提高SNR。与其他独立于说话者和基于模型的方法相比,该系统具有竞争优势。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号