首页> 外文会议>INTERSPEECH 2012 >Discriminatively learning factorized finite state pronunciation models from dynamic Bayesian networks
【24h】

Discriminatively learning factorized finite state pronunciation models from dynamic Bayesian networks

机译:差异地学习动态贝叶斯网络的有限状态发音模型

获取原文

摘要

This paper describes an approach to efficiently construct, and discriminatively train,a weighted finite state transducer (WFST) representation for an articulatory feature-based model of pronunciation. This model is originally implemented as a dynamic Bayesian network (DBN). The Work is motivated by a desire to (1) incorporate such a pronunciation model in WFST-based recognizers, and to (2) learn discriminative models that are more generalthan the DBNs. The approach is quite general, though here we show how it applies to a specific model. We use the conditional independence assumptions imposed by the DBN to efficiently convert it into a sequence of WFSTs (factor FSTs) which, when composed, yield the. same model as the DBN. We then introduce a linear model of the arc weights of the factor FSTs arid discriminatively learn its weights using the averaged perceptron algorithm. We demonstrate the approach using a lexical access task in which we recognize a word given its surface realization. Our experimental results using a phonetically transcribed subset of the Switchboard corpus show that the discriminatively learned model performs significantly better than the original DBN.
机译:本文描述的方法以有效地构建体,和区别地火车,用于发音的关节基于特征的模型的加权有限状态变换器(WFST)表示。这种模式最初实现为动态贝叶斯网络(DBN)。这项工作是由一个愿望所驱使:(1)结合基于WFST的识别这样一个发音模型,以及(2)学会判别模型是更generalthan的动态Bayesian。该方法是相当普遍的,但在这里我们展示它如何适用于特定的模式。我们使用由DBN规定的条件独立性假设有效地将其转换成WFSTs(因子FSTS),其组成时,得到的序列。同一型号的DBN。然后,我们介绍因素FSTS的弧权重的线性模型使用平均感知算法干旱有区别地了解它的权重。我们演示使用中,我们认识到由于其表面实现一个字一个词接任务的方式。我们使用音素转录的总机语料库秀子集实验结果表明,显著优于原DBN的区别性学习模型执行。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号