【24h】

Automatic Language Identification Using Mixed-order Hmms And Untranscribed Corpora

机译:使用混合顺序Hmm和未转录语料库的自动语言识别

获取原文

摘要

The state-of-the-art language identification (LID) systems are based on phone recognisers and n-gram language models, which require the use of transcribed speech databases for training. An alternate solution to the LID problem directly applies mixed-order hidden Markov models (HMMs) to untranscribed speech. The competitive performance of these mixed-order HMMs on the NIST 1996 evaluation set is very promising, considering the ease of implementation and many possible improvements. This validates a novel mixed-order HMM training procedure and extends previous results obtained with high-order HMMs to take advantage of larger datasets.
机译:最新的语言识别(LID)系统基于电话识别器和n-gram语言模型,这需要使用转录语音数据库进行培训。 LID问题的另一种解决方案将混合顺序隐马尔可夫模型(HMM)直接应用于非转录语音。考虑到易于实施和许多可能的改进,这些混合顺序HMM在NIST 1996评估套件上的竞争性能非常有前途。这验证了一种新颖的混合阶HMM训练程序,并扩展了使用高阶HMM获得的先前结果,以利用更大的数据集。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号