首页> 外文会议>Workshop on Automatic Speech Recognition and Understanding >K-COMPONENT RECURRENT NEURAL NETWORK LANGUAGE MODELS USING CURRICULUM LEARNING
【24h】

K-COMPONENT RECURRENT NEURAL NETWORK LANGUAGE MODELS USING CURRICULUM LEARNING

机译:K-Component经常性的神经网络语言模型使用课程学习

获取原文

摘要

Conventional n-gram language models are known for their limited ability to capture long-distance dependencies and their brittleness with respect to within-domain variations. In this paper, we propose a k-component recurrent neural network language model using curriculum learning (CL-krnnlm) to address within-domain variations. Based on a Dutch-language corpus, we investigate three methods of curriculum learning that exploit dedicated component models for specific sub-domains. Under an oracle situation in which context information is known during testing, we experimentally test three hypotheses. The first is that domain-dedicated models perform better than general models on their specific domains. The second is that curriculum learning can be used to train recurrent neural network language models (RNNLMs) from general patterns to specific patterns. The third is that curriculum learning, used as an implicit weighting method to adjust the relative contributions of general and specific patterns, outperforms conventional linear interpolation. Under the condition that context information is unknown during testing, the cl-krnnlm also achieves improvement over conventional RNNLM by 13% relative in terms of word prediction accuracy. Finally, the CL-KRNNLM is tested in an additional experiment involving N-best rescoring on a standard data set. Here, the context domains are created by clustering the training data using Latent Dirichlet Allocation and k-means clustering.
机译:已知传统的N-GRAM语言模型以其有限的能力来捕获长距离依赖性及其脆性相对于域内变化的能力。在本文中,我们提出了使用课程学习(CL-KRNNLM)来解决域内变化的K-Component复发性神经网络语言模型。基于荷兰语语料库,我们调查了三种课程学习方法,用于专用特定子域的专用组件模型。根据在测试期间已知上下文信息的oracle情况下,我们通过实验测试三个假设。首先是域专用模型比其特定域上的一般模型更好地执行。其次是,课程学习可用于将经常性神经网络语言模型(RNNLMS)从一般模式培训到特定模式。第三是课程学习,用作一种隐式加权方法来调整一般和特定模式的相对贡献,优于传统的线性插值。在测试期间上下文信息未知的条件下,CL-KrnNLM在字形预测精度方面,CL-KrnNLM还通过13%的传统RNNLM实现了改进。最后,在涉及在标准数据集的N-Best Creving的附加实验中测试CL-KrnNLM。这里,通过使用潜在的Dirichlet分配和K-means群集群集培训数据来创建上下文域。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号