首页> 外文期刊>IEICE Transactions on Information and Systems >Bayesian Learning of a Language Model from Continuous Speech
【24h】

Bayesian Learning of a Language Model from Continuous Speech

机译:从连续语音中进行语言模型的贝叶斯学习

获取原文
获取原文并翻译 | 示例
       

摘要

We propose a novel scheme to learn a language model (LM) for automatic speech recognition (ASR) directly from continuous speech. In the proposed method, we first generate phoneme lattices using an acoustic model with no linguistic constraints, then perform training over these phoneme lattices, simultaneously learning both lexical units and an LM. As a statistical framework for this learning problem, we use non-parametric Bayesian statistics, which make it possible to balance the learned model's complexity (such as the size of the learned vocabulary) and expressive power, and provide a principled learning algorithm through the use of Gibbs sampling. Implementation is performed using weighted finite state transducers (WFSTs), which allow for the simple handling of lattice input. Experimental results on natural, adult-directed speech demonstrate that LMs built using only continuous speech are able to significantly reduce ASR phoneme error rates. The proposed technique of joint Bayesian learning of lexical units and an LM over lattices is shown to significantly contribute to this improvement.
机译:我们提出了一种新颖的方案,可以直接从连续语音中学习用于自动语音识别(ASR)的语言模型(LM)。在提出的方法中,我们首先使用没有语言约束的声学模型生成音素格,然后对这些音素格进行训练,同时学习词汇单元和LM。作为此学习问题的统计框架,我们使用非参数贝叶斯统计方法,可以平衡学习模型的复杂性(例如学习词汇的大小)和表达能力,并通过使用提供有原则的学习算法Gibbs抽样。使用加权有限状态换能器(WFST)来执行实现,这允许简单地处理晶格输入。关于自然的,成人定向语音的实验结果表明,仅使用连续语音构建的LM能够显着降低ASR音素错误率。提出的联合贝叶斯学习词法单元和基于矩阵的LM的建议技术将显着促进这一改进。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号