首页> 外文会议>2017 IEEE Automatic Speech Recognition and Understanding Workshop >Feature optimized DPGMM clustering for unsupervised subword modeling: A contribution to zerospeech 2017
【24h】

Feature optimized DPGMM clustering for unsupervised subword modeling: A contribution to zerospeech 2017

机译:针对无监督子词建模的功能优化的DPGMM聚类:对Zerospeech 2017的贡献

获取原文
获取原文并翻译 | 示例

摘要

This paper describes our unsupervised subword modeling pipeline for the zero resource speech challenge (ZeroSpeech) 2017. Our approach is built around the Dirichlet process Gaussian mixture model (DPGMM) that we use to cluster speech feature vectors into a dynamically sized set of classes. By considering each class an acoustic unit, speech can be represented as sequence of class posteriorgrams. We enhance this method by automatically optimizing the DPGMM sampler's input features in a multi-stage clustering framework, where we unsupervisedly learn transformations using LDA, MLLT and (basis) fMLLR to reduce variance in the features. We show that this optimization considerably boosts the subword modeling quality, according to the performance on the ABX phone discriminability task. For the first time, we apply inferred subword models to previously unseen data from a new set of speakers. We demonstrate our method's good generalization and the effectiveness of its blind speaker adaptation in extensive experiments on a multitude of datasets. Our pipeline has very little need for hyper-parameter adjustment and is entirely unsupervised, i.e., it only takes raw audio recordings as input, without requiring any pre-defined segmentation, explicit speaker IDs or other meta data.
机译:本文介绍了针对零资源语音挑战(ZeroSpeech)2017的无监督子词建模管道。我们的方法是围绕Dirichlet过程高斯混合模型(DPGMM)构建的,该模型用于将语音特征向量聚类为动态大小的类集。通过将每个类别考虑为一个声学单位,语音可以表示为一系列后验图。我们通过在多阶段聚类框架中自动优化DPGMM采样器的输入特征来增强此方法,在该框架中,我们无监督地学习了使用LDA,MLLT和(基本)fMLLR的变换,以减少特征的差异。我们显示,根据ABX电话可分辨性任务上的性能,此优化可以大大提高子词建模质量。第一次,我们将推断的子词模型应用于来自一组新说话者的先前看不见的数据。我们在大量数据集的广泛实验中证明了该方法的良好概括性以及盲人适应性的有效性。我们的管道几乎不需要超参数调整,并且完全不受监督,即仅将原始音频记录作为输入,而无需任何预定义的分段,明确的说话者ID或其他元数据。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号