首页> 外文会议>Annual conference of the International Speech Communication Association >Discriminatively learning factorized finite state pronunciation models from dynamic Bayesian networks
【24h】

Discriminatively learning factorized finite state pronunciation models from dynamic Bayesian networks

机译:从动态贝叶斯网络区别学习因式分解有限状态发音模型

获取原文

摘要

This paper describes an approach to efficiently construct, and discriminatively train, a weighted finite state transducer (WFST) representation for an articulatory feature-based model of pronunciation. This model is originally implemented as a dynamic Bayesian network (DBN). The work is motivated by a desire to (1) incorporate such a pronunciation model in WFST-based recognizers, and to (2) learn discriminative models that are more general than the DBNs. The approach is quite general, though here we show how it applies to a specific model. We use the conditional independence assumptions imposed by the DBN to efficiently convert it into a sequence of WFSTs (factor FSTs) which, when composed, yield the same model as the DBN. We then introduce a linear model of the arc weights of the factor FSTs and discriminatively learn its weights using the averaged perceptron algorithm. We demonstrate the approach using a lexical access task in which we recognize a word given its surface realization. Our experimental results using a phonetically transcribed subset of the Switchboard corpus show that the discriminatively learned model performs significantly better than the original DBN.
机译:本文介绍了一种方法,该方法可有效构造和区分训练基于发音特征的发音模型的加权有限状态换能器(WFST)表示。该模型最初实现为动态贝叶斯网络(DBN)。这项工作的动机是希望(1)将这种语音模型纳入基于WFST的识别器中,以及(2)学习比DBN更通用的判别模型。该方法相当笼统,尽管在这里我们展示了它如何应用于特定模型。我们使用DBN施加的条件独立性假设,将其有效地转换为一系列WFST(因子FST),这些序列组成时会产生与DBN相同的模型。然后,我们引入因子FST的弧权重的线性模型,并使用平均感知器算法来区别地学习其权重。我们使用词法访问任务演示了该方法,在该方法中,我们根据表面实现识别了一个单词。我们使用Switchboard语料库的语音转录子集进行的实验结果表明,有区别学习的模型的性能明显优于原始DBN。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号