首页> 外文期刊>IEEE Transactions on Signal Processing >Phonemic hidden Markov models with continuous mixture output densities for large vocabulary word recognition
【24h】

Phonemic hidden Markov models with continuous mixture output densities for large vocabulary word recognition

机译:具有连续混合输出密度的音素隐马尔可夫模型,可用于大词汇量单词识别

获取原文
获取原文并翻译 | 示例
获取外文期刊封面目录资料

摘要

The authors demonstrate the effectiveness of phonemic hidden Markov models with Gaussian mixture output densities (mixture HMMs) for speaker-dependent large-vocabulary word recognition. Speech recognition experiments show that for almost any reasonable amount of training data, recognizers using mixture HMMs consistently outperform those employing unimodal Gaussian HMMs. With a sufficiently large training set (e.g. more than 2500 words), use of HMMs with 25-component mixture distributions typically reduces recognition errors by about 40%. It is also found that the mixture HMMs outperform a set of unimodal generalized triphone models having the same number of parameters. Previous attempts to employ mixture HMMs for speech recognition proved discouraging because of the high complexity and computational cost in implementing the Baum-Welch training algorithm. It is shown how mixture HMMs can be implemented very simply in unimodal transition-based frameworks by allowing multiple transitions from one state to another.
机译:作者展示了具有高斯混合输出密度(混合HMM)的语音隐马尔可夫模型对于说话者相关的大词汇量单词识别的有效性。语音识别实验表明,对于几乎任何合理数量的训练数据,使用混合HMM的识别器始终优于使用单峰高斯HMM的识别器。在足够大的训练集(例如超过2500个单词)的情况下,使用具有25种成分的混合物分布的HMM通常可以将识别错误减少约40%。还发现混合HMM优于一组具有相同数量参数的单峰广义三音器模型。由于采用Baum-Welch训练算法的复杂性和计算成本较高,以前采用混合HMM进行语音识别的尝试令人沮丧。它显示了如何通过允许从一种状态到另一种状态的多次转换,在基于单峰转换的框架中非常简单地实现混合HMM。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号