首页> 外文会议>Annual conference of the International Speech Communication Association;INTERSPEECH 2011 >Effective Triphone Mapping for Acoustic Modeling in Speech Recognition
【24h】

Effective Triphone Mapping for Acoustic Modeling in Speech Recognition

机译:用于语音识别中声学建模的有效Triphone映射

获取原文

摘要

This paper presents effective triphone mapping for acoustic models training in automatic speech recognition, which allows the synthesis of unseen triphones. The description of this data-driven model clustering, including experiments performed using 350 hours of a Slovak audio database of mixed read and spontaneous speech, are presented. The proposed technique is compared with tree-based state tying, and it is shown that for bigger acoustic models, at a size of 4000 states and more, a triphone mapped HMM system achieves better performance than a tree-based state tying system. The main gain in performance is due to latent application of triphone mapping on monophones with multiple Gaussian pdfs, so the cloned triphones are initialized better than with single Gaussians monophones. Absolute decrease of word error rate was 0.46% (5.73% relatively) for models with 7500 states, and decreased to 0.4% (5.17% relatively) gain at 11500 states.
机译:本文介绍了有效的三音映射,用于自动语音识别中的声学模型训练,从而可以合成看不见的三音。介绍了这种数据驱动的模型聚类的描述,包括使用350小时的混合阅读和自发语音的Slovak音频数据库进行的实验。将该技术与基于树的状态绑定系统进行了比较,结果表明,对于较大的声学模型,在4000个状态以及更大的状态下,三音机映射的HMM系统比基于树的状态绑定系统具有更好的性能。性能的主要提高是由于在具有多个高斯pdf的单音手机上潜在地应用了三音手机映射,因此,与使用单个高斯单音手机相比,克隆的三音手机的初始化效果更好。在7500个州的模型中,单词错误率的绝对降低为0.46%(相对为5.73%),在11500个州中降至0.4%(相对为5.17%)。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号