首页> 外文会议>Annual conference of the International Speech Communication Association >Towards Recurrent Neural Networks Language Models with Linguistic and Contextual Features
【24h】

Towards Recurrent Neural Networks Language Models with Linguistic and Contextual Features

机译:具有语言和上下文特征的递归神经网络语言模型

获取原文

摘要

Recent studies show that recurrent neural network language models (RNNLM) perform better than traditional language models such as smoothed n-grams. For traditional models it is known that the addition of for example part-of-speech information and topical information can improve performance. In this paper we investigate the usefulness of additional features for RNNLM. We look at four types of features: POS tags, lemmas, and the topics and the socio-situational setting of a conversation. In our experiments, almost all RNNLM models that make use of extra information outperform our baseline RNNLM model in terms of both perplexity and word prediction accuracy. Whereas the baseline model has a perplexity of 114.79, the model that uses a combination of POS tags, socio-situational settings and lemmas achieves the lowest perplexity result of 83.59, and the combination of all 4 types of features, using a network with 500 hidden neurons, achieves the highest word prediction accuracy of 23.11%.
机译:最近的研究表明,递归神经网络语言模型(RNNLM)的性能比传统语言模型(如平滑n-gram)好。对于传统模型,众所周知,添加词性信息和主题信息可以提高性能。在本文中,我们研究了附加功能对RNNLM的有用性。我们研究四种类型的功能:POS标签,引理以及话题和对话的社会情境设置。在我们的实验中,几乎所有使用额外信息的RNNLM模型在困惑度和单词预测准确性方面都优于我们的基线RNNLM模型。基准模型的困惑度为114.79,而结合使用POS标签,社会情境设置和引理的模型,则使用500个隐藏的网络实现了最低的困惑度结果83.59,以及所有4种特征的组合。神经元,达到23.11%的最高单词预测准确率。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号