【24h】

Reducing infrequent-token perplexity via variational corpora

机译:通过变体语料减少不频繁的令牌困惑

获取原文

摘要

Recurrent neural network (RNN) is recognized as a powerful language model (LM). We investigate deeper into its performance portfolio, which performs well on frequent grammatical patterns but much less so on less frequent terms. Such portfolio is expected and desirable in applications like autocomplete, but is less useful in social content analysis where many creative, unexpected usages occur (e.g., URL insertion). We adapt a generic RNN model and show that, with variational training corpora and epoch unfolding, the model improves its performance for the task of URL insertion suggestions.
机译:递归神经网络(RNN)被认为是功能强大的语言模型(LM)。我们对其性能组合进行了更深入的研究,该组合在常见的语法模式下表现良好,而在较不频繁的条件下则表现不佳。在诸如自动完成之类的应用中,这种组合是期望的并且是期望的,但是在社交内容分析中却没有多大用处,在社交内容分析中会发生许多创造性的意外使用(例如URL插入)。我们改编了通用RNN模型,并显示出,通过变体训练语料库和历元展开,该模型提高了其针对URL插入建议任务的性能。

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号